var/home/core/zuul-output/0000755000175000017500000000000015073464033014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015073471105015474 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004371501215073471077017713 0ustar rootrootOct 14 15:14:28 crc systemd[1]: Starting Kubernetes Kubelet... Oct 14 15:14:28 crc restorecon[4669]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:28 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 15:14:29 crc restorecon[4669]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 15:14:29 crc restorecon[4669]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 14 15:14:30 crc kubenswrapper[4945]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 15:14:30 crc kubenswrapper[4945]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 14 15:14:30 crc kubenswrapper[4945]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 15:14:30 crc kubenswrapper[4945]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 15:14:30 crc kubenswrapper[4945]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 14 15:14:30 crc kubenswrapper[4945]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.477490 4945 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.482986 4945 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483017 4945 feature_gate.go:330] unrecognized feature gate: Example Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483029 4945 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483039 4945 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483048 4945 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483056 4945 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483079 4945 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483088 4945 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483095 4945 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483103 4945 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483114 4945 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483124 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483134 4945 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483143 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483151 4945 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483159 4945 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483166 4945 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483174 4945 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483182 4945 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483190 4945 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483198 4945 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483206 4945 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483214 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483223 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483231 4945 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483239 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483246 4945 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483254 4945 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483263 4945 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483271 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483279 4945 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483286 4945 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483294 4945 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483302 4945 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483309 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483317 4945 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483325 4945 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483334 4945 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483343 4945 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483350 4945 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483359 4945 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483369 4945 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483379 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483387 4945 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483395 4945 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483402 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483410 4945 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483422 4945 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483432 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483442 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483450 4945 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483460 4945 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483468 4945 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483476 4945 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483485 4945 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483493 4945 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483501 4945 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483509 4945 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483517 4945 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483525 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483533 4945 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483540 4945 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483548 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483556 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483564 4945 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483572 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483580 4945 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483587 4945 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483604 4945 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483616 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.483626 4945 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483804 4945 flags.go:64] FLAG: --address="0.0.0.0" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483822 4945 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483837 4945 flags.go:64] FLAG: --anonymous-auth="true" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483849 4945 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483861 4945 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483871 4945 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483908 4945 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483918 4945 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483928 4945 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483937 4945 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483947 4945 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483956 4945 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483965 4945 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483974 4945 flags.go:64] FLAG: --cgroup-root="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483983 4945 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.483992 4945 flags.go:64] FLAG: --client-ca-file="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484000 4945 flags.go:64] FLAG: --cloud-config="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484009 4945 flags.go:64] FLAG: --cloud-provider="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484017 4945 flags.go:64] FLAG: --cluster-dns="[]" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484030 4945 flags.go:64] FLAG: --cluster-domain="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484038 4945 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484048 4945 flags.go:64] FLAG: --config-dir="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484056 4945 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484066 4945 flags.go:64] FLAG: --container-log-max-files="5" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484077 4945 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484086 4945 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484095 4945 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484104 4945 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484112 4945 flags.go:64] FLAG: --contention-profiling="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484125 4945 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484135 4945 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484145 4945 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484155 4945 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484166 4945 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484175 4945 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484185 4945 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484193 4945 flags.go:64] FLAG: --enable-load-reader="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484202 4945 flags.go:64] FLAG: --enable-server="true" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484211 4945 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484221 4945 flags.go:64] FLAG: --event-burst="100" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484230 4945 flags.go:64] FLAG: --event-qps="50" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484240 4945 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484249 4945 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484258 4945 flags.go:64] FLAG: --eviction-hard="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484268 4945 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484277 4945 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484286 4945 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484296 4945 flags.go:64] FLAG: --eviction-soft="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484304 4945 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484313 4945 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484323 4945 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484332 4945 flags.go:64] FLAG: --experimental-mounter-path="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484340 4945 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484349 4945 flags.go:64] FLAG: --fail-swap-on="true" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484358 4945 flags.go:64] FLAG: --feature-gates="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484368 4945 flags.go:64] FLAG: --file-check-frequency="20s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484377 4945 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484387 4945 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484396 4945 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484405 4945 flags.go:64] FLAG: --healthz-port="10248" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484414 4945 flags.go:64] FLAG: --help="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484426 4945 flags.go:64] FLAG: --hostname-override="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484435 4945 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484444 4945 flags.go:64] FLAG: --http-check-frequency="20s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484453 4945 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484462 4945 flags.go:64] FLAG: --image-credential-provider-config="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484471 4945 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484480 4945 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484491 4945 flags.go:64] FLAG: --image-service-endpoint="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484500 4945 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484509 4945 flags.go:64] FLAG: --kube-api-burst="100" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484518 4945 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484527 4945 flags.go:64] FLAG: --kube-api-qps="50" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484536 4945 flags.go:64] FLAG: --kube-reserved="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484545 4945 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484553 4945 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484562 4945 flags.go:64] FLAG: --kubelet-cgroups="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484571 4945 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484580 4945 flags.go:64] FLAG: --lock-file="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484588 4945 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484597 4945 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484606 4945 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484619 4945 flags.go:64] FLAG: --log-json-split-stream="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484628 4945 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484637 4945 flags.go:64] FLAG: --log-text-split-stream="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484646 4945 flags.go:64] FLAG: --logging-format="text" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484655 4945 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484664 4945 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484673 4945 flags.go:64] FLAG: --manifest-url="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484682 4945 flags.go:64] FLAG: --manifest-url-header="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484693 4945 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484702 4945 flags.go:64] FLAG: --max-open-files="1000000" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484713 4945 flags.go:64] FLAG: --max-pods="110" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484726 4945 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484736 4945 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484745 4945 flags.go:64] FLAG: --memory-manager-policy="None" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484754 4945 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484763 4945 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484773 4945 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484781 4945 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484800 4945 flags.go:64] FLAG: --node-status-max-images="50" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484809 4945 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484818 4945 flags.go:64] FLAG: --oom-score-adj="-999" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484827 4945 flags.go:64] FLAG: --pod-cidr="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484837 4945 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484849 4945 flags.go:64] FLAG: --pod-manifest-path="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484858 4945 flags.go:64] FLAG: --pod-max-pids="-1" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484867 4945 flags.go:64] FLAG: --pods-per-core="0" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484901 4945 flags.go:64] FLAG: --port="10250" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484913 4945 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484923 4945 flags.go:64] FLAG: --provider-id="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484934 4945 flags.go:64] FLAG: --qos-reserved="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484945 4945 flags.go:64] FLAG: --read-only-port="10255" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484957 4945 flags.go:64] FLAG: --register-node="true" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484967 4945 flags.go:64] FLAG: --register-schedulable="true" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484975 4945 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.484990 4945 flags.go:64] FLAG: --registry-burst="10" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485000 4945 flags.go:64] FLAG: --registry-qps="5" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485009 4945 flags.go:64] FLAG: --reserved-cpus="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485018 4945 flags.go:64] FLAG: --reserved-memory="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485028 4945 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485037 4945 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485046 4945 flags.go:64] FLAG: --rotate-certificates="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485055 4945 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485064 4945 flags.go:64] FLAG: --runonce="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485077 4945 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485086 4945 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485096 4945 flags.go:64] FLAG: --seccomp-default="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485105 4945 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485114 4945 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485123 4945 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485132 4945 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485141 4945 flags.go:64] FLAG: --storage-driver-password="root" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485149 4945 flags.go:64] FLAG: --storage-driver-secure="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485181 4945 flags.go:64] FLAG: --storage-driver-table="stats" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485192 4945 flags.go:64] FLAG: --storage-driver-user="root" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485222 4945 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485235 4945 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485246 4945 flags.go:64] FLAG: --system-cgroups="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485257 4945 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485295 4945 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485306 4945 flags.go:64] FLAG: --tls-cert-file="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485317 4945 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485338 4945 flags.go:64] FLAG: --tls-min-version="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485349 4945 flags.go:64] FLAG: --tls-private-key-file="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485361 4945 flags.go:64] FLAG: --topology-manager-policy="none" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485371 4945 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485382 4945 flags.go:64] FLAG: --topology-manager-scope="container" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485393 4945 flags.go:64] FLAG: --v="2" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485406 4945 flags.go:64] FLAG: --version="false" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485421 4945 flags.go:64] FLAG: --vmodule="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485433 4945 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.485445 4945 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485668 4945 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485678 4945 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485687 4945 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485695 4945 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485706 4945 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485716 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485724 4945 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485733 4945 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485741 4945 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485748 4945 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485756 4945 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485763 4945 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485772 4945 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485779 4945 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485787 4945 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485794 4945 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485802 4945 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485809 4945 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485818 4945 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485825 4945 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485833 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485840 4945 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485848 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485857 4945 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485864 4945 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485898 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485906 4945 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485917 4945 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485927 4945 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485936 4945 feature_gate.go:330] unrecognized feature gate: Example Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485945 4945 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485953 4945 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485962 4945 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485971 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485979 4945 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485988 4945 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.485999 4945 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486007 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486015 4945 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486023 4945 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486030 4945 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486038 4945 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486046 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486054 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486064 4945 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486073 4945 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486082 4945 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486090 4945 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486098 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486106 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486114 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486124 4945 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486133 4945 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486141 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486149 4945 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486156 4945 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486164 4945 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486173 4945 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486181 4945 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486191 4945 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486198 4945 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486206 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486214 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486221 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486229 4945 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486237 4945 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486244 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486251 4945 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486263 4945 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486271 4945 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.486279 4945 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.486998 4945 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.499735 4945 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.499790 4945 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.499948 4945 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.499970 4945 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.499983 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.499994 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500004 4945 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500013 4945 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500022 4945 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500033 4945 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500047 4945 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500057 4945 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500068 4945 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500079 4945 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500091 4945 feature_gate.go:330] unrecognized feature gate: Example Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500100 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500110 4945 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500119 4945 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500128 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500138 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500150 4945 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500160 4945 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500168 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500177 4945 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500185 4945 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500194 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500202 4945 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500210 4945 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500219 4945 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500227 4945 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500236 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500244 4945 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500252 4945 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500261 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500269 4945 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500278 4945 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500288 4945 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500296 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500305 4945 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500313 4945 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500321 4945 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500330 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500338 4945 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500346 4945 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500354 4945 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500363 4945 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500371 4945 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500379 4945 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500387 4945 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500396 4945 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500405 4945 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500413 4945 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500421 4945 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500429 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500437 4945 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500446 4945 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500454 4945 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500465 4945 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500476 4945 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500485 4945 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500495 4945 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500505 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500514 4945 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500523 4945 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500532 4945 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500541 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500550 4945 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500558 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500566 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500574 4945 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500582 4945 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500591 4945 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500600 4945 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.500614 4945 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500855 4945 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500870 4945 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500903 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500912 4945 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500922 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500930 4945 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500939 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500948 4945 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500956 4945 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500966 4945 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500978 4945 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.500990 4945 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501000 4945 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501010 4945 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501019 4945 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501029 4945 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501038 4945 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501047 4945 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501056 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501064 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501072 4945 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501080 4945 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501088 4945 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501097 4945 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501105 4945 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501114 4945 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501124 4945 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501134 4945 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501143 4945 feature_gate.go:330] unrecognized feature gate: Example Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501174 4945 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501186 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501197 4945 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501206 4945 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501216 4945 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501226 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501235 4945 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501244 4945 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501252 4945 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501265 4945 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501276 4945 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501286 4945 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501296 4945 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501307 4945 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501316 4945 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501326 4945 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501335 4945 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501346 4945 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501356 4945 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501364 4945 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501373 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501381 4945 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501392 4945 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501400 4945 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501408 4945 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501417 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501425 4945 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501433 4945 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501441 4945 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501450 4945 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501459 4945 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501467 4945 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501477 4945 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501486 4945 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501494 4945 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501503 4945 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501511 4945 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501519 4945 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501528 4945 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501536 4945 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501545 4945 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.501555 4945 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.501568 4945 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.502773 4945 server.go:940] "Client rotation is on, will bootstrap in background" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.508862 4945 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.509070 4945 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.511031 4945 server.go:997] "Starting client certificate rotation" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.511091 4945 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.512201 4945 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-28 19:49:59.759112916 +0000 UTC Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.512308 4945 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1804h35m29.246809489s for next certificate rotation Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.545288 4945 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.550662 4945 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.575222 4945 log.go:25] "Validated CRI v1 runtime API" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.617160 4945 log.go:25] "Validated CRI v1 image API" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.619653 4945 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.627806 4945 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-14-15-09-21-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.627921 4945 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.657159 4945 manager.go:217] Machine: {Timestamp:2025-10-14 15:14:30.653243366 +0000 UTC m=+0.637291794 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:6a76cd5d-bafd-45c2-978d-e5848e35a4cd BootID:d936a6ed-46ff-4fc3-bc0a-62bc04b87659 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:87:ca:db Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:87:ca:db Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:65:e5:bc Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:6e:bc:7d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:4d:7c:ac Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e6:e2:18 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:86:b5:7e:7c:a7:d6 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:aa:8b:db:6b:06:a2 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.657564 4945 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.657783 4945 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.659397 4945 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.659735 4945 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.659793 4945 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.660191 4945 topology_manager.go:138] "Creating topology manager with none policy" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.660210 4945 container_manager_linux.go:303] "Creating device plugin manager" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.660631 4945 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.660680 4945 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.663802 4945 state_mem.go:36] "Initialized new in-memory state store" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.664076 4945 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.673853 4945 kubelet.go:418] "Attempting to sync node with API server" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.673938 4945 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.674011 4945 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.674035 4945 kubelet.go:324] "Adding apiserver pod source" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.674063 4945 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.679530 4945 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.681338 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.681345 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:30 crc kubenswrapper[4945]: E1014 15:14:30.681515 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.682059 4945 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 14 15:14:30 crc kubenswrapper[4945]: E1014 15:14:30.682190 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.686052 4945 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689600 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689649 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689664 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689678 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689701 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689716 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689730 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689752 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689768 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689784 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689802 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689817 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.689847 4945 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.690680 4945 server.go:1280] "Started kubelet" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.690826 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.692007 4945 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.692022 4945 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.692814 4945 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 14 15:14:30 crc systemd[1]: Started Kubernetes Kubelet. Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.694868 4945 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.694956 4945 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.695321 4945 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 04:41:48.034749512 +0000 UTC Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.695384 4945 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1045h27m17.339369881s for next certificate rotation Oct 14 15:14:30 crc kubenswrapper[4945]: E1014 15:14:30.695395 4945 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.695442 4945 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.695456 4945 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.695557 4945 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.696563 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:30 crc kubenswrapper[4945]: E1014 15:14:30.696676 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.696805 4945 factory.go:55] Registering systemd factory Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.697244 4945 server.go:460] "Adding debug handlers to kubelet server" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.697747 4945 factory.go:221] Registration of the systemd container factory successfully Oct 14 15:14:30 crc kubenswrapper[4945]: E1014 15:14:30.697909 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.245:6443: connect: connection refused" interval="200ms" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.699411 4945 factory.go:153] Registering CRI-O factory Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.699561 4945 factory.go:221] Registration of the crio container factory successfully Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.699800 4945 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.699969 4945 factory.go:103] Registering Raw factory Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.700077 4945 manager.go:1196] Started watching for new ooms in manager Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.700904 4945 manager.go:319] Starting recovery of all containers Oct 14 15:14:30 crc kubenswrapper[4945]: E1014 15:14:30.700264 4945 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.245:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186e6460aa9f4f3b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-14 15:14:30.690647867 +0000 UTC m=+0.674696275,LastTimestamp:2025-10-14 15:14:30.690647867 +0000 UTC m=+0.674696275,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.712177 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.712459 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.712583 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.712694 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.712828 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.712971 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.713127 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.713240 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.713366 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.713475 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.713593 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.713699 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.713844 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.714004 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.714126 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.714234 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.714355 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.714471 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.714576 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.714716 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.714829 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.714974 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.715095 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.715203 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.715316 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.715422 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.715547 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.715675 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.716477 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.716614 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.716720 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.716834 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.716985 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.717141 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.717256 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.717366 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.717468 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.717569 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.717678 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.717793 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.717917 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.718024 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.718134 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.718255 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.718363 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.718467 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.718571 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.718684 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.718796 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.718947 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.719070 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.719178 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.719283 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.719407 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.719551 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.719669 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.719783 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.719941 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.720066 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.720203 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.720322 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.720435 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.720545 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.720766 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.720927 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.721087 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.721219 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.721339 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.721451 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.723746 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.723793 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.723840 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.723854 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.723867 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.723924 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.723938 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.723950 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.723962 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.723978 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724001 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724016 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724031 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724045 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724060 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724073 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724087 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724099 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724111 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724123 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724135 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724149 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724164 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724226 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724241 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724253 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724271 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724284 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724299 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724313 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724328 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724350 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724363 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724376 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724390 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724413 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724431 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724446 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724461 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724477 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724493 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724508 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724525 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724540 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724555 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724569 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724582 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724596 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724609 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724624 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724639 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724651 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724664 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724678 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724692 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724705 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724718 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724731 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724747 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724761 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724774 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724789 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724803 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724816 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724830 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724845 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724858 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724889 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724903 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724916 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724928 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724941 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724953 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724968 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724982 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.724995 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725007 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725020 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725034 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725076 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725088 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725102 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725115 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725130 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725144 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725157 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725171 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725184 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725197 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725210 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725225 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725238 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725249 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725263 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725277 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725291 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725303 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725317 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725331 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725344 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725358 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725371 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725384 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725397 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725409 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725422 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725435 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725448 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725461 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725475 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725488 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725533 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725547 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725561 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725574 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725597 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725611 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725624 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725637 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725650 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725689 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725701 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725720 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725733 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725745 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725759 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725773 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725786 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725799 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725812 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725826 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725838 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725850 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.725864 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.727587 4945 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.727619 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.727638 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.727656 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.727676 4945 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.727689 4945 reconstruct.go:97] "Volume reconstruction finished" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.727698 4945 reconciler.go:26] "Reconciler: start to sync state" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.733297 4945 manager.go:324] Recovery completed Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.741610 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.743001 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.743051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.743063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.743959 4945 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.743987 4945 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.744015 4945 state_mem.go:36] "Initialized new in-memory state store" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.759217 4945 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.759845 4945 policy_none.go:49] "None policy: Start" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.761315 4945 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.761369 4945 state_mem.go:35] "Initializing new in-memory state store" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.761576 4945 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.761623 4945 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.761654 4945 kubelet.go:2335] "Starting kubelet main sync loop" Oct 14 15:14:30 crc kubenswrapper[4945]: E1014 15:14:30.761704 4945 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 14 15:14:30 crc kubenswrapper[4945]: W1014 15:14:30.763401 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:30 crc kubenswrapper[4945]: E1014 15:14:30.763484 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 14 15:14:30 crc kubenswrapper[4945]: E1014 15:14:30.795622 4945 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.822821 4945 manager.go:334] "Starting Device Plugin manager" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.822900 4945 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.822915 4945 server.go:79] "Starting device plugin registration server" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.823328 4945 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.823346 4945 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.823524 4945 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.823613 4945 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.823623 4945 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 14 15:14:30 crc kubenswrapper[4945]: E1014 15:14:30.830021 4945 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.861773 4945 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.861916 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.863114 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.863181 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.863193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.863401 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.863663 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.863734 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.864416 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.864446 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.864454 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.864564 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.864739 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.864780 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.864892 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.864913 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.864921 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.865608 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.865642 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.865658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.865848 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.865867 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.865907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.865920 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.866132 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.866169 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.866838 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.866846 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.866861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.866891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.866898 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.866907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.867045 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.867209 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.867258 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.867913 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.867944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.867955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.868115 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.868145 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.868609 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.868631 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.868640 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.868803 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.868826 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.868836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:30 crc kubenswrapper[4945]: E1014 15:14:30.898550 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.245:6443: connect: connection refused" interval="400ms" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.923734 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.925079 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.925118 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.925127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.925157 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 15:14:30 crc kubenswrapper[4945]: E1014 15:14:30.925556 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.245:6443: connect: connection refused" node="crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.929908 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.929944 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.929963 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.929981 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.930000 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.930015 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.930029 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.930042 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.930058 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.930071 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.930085 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.930099 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.930115 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.930133 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:30 crc kubenswrapper[4945]: I1014 15:14:30.930150 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.031678 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.031746 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.031783 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.031814 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.031840 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.031867 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.031915 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.031945 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.031970 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.031997 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032022 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032049 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032324 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032187 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032356 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032396 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032422 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032436 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032501 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032242 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032279 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032286 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032199 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032213 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032233 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032438 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032238 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032460 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032242 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.032257 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.125915 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.127393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.127430 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.127442 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.127466 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 15:14:31 crc kubenswrapper[4945]: E1014 15:14:31.127909 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.245:6443: connect: connection refused" node="crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.187361 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.196273 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.219585 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.232907 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.240548 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 15:14:31 crc kubenswrapper[4945]: W1014 15:14:31.242006 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-60092bc1686735e72e267ef3854e7d6725316aae38b943aed170058bb174e6fe WatchSource:0}: Error finding container 60092bc1686735e72e267ef3854e7d6725316aae38b943aed170058bb174e6fe: Status 404 returned error can't find the container with id 60092bc1686735e72e267ef3854e7d6725316aae38b943aed170058bb174e6fe Oct 14 15:14:31 crc kubenswrapper[4945]: W1014 15:14:31.242603 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-2b9270bc97ef89ed5a7cfec7778208e7401791587988a126a96736b95b7266d5 WatchSource:0}: Error finding container 2b9270bc97ef89ed5a7cfec7778208e7401791587988a126a96736b95b7266d5: Status 404 returned error can't find the container with id 2b9270bc97ef89ed5a7cfec7778208e7401791587988a126a96736b95b7266d5 Oct 14 15:14:31 crc kubenswrapper[4945]: W1014 15:14:31.256031 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-3d065435931ba0372ae31bb912a5e6a9655979c289b202c89eb5dbc2079f71d9 WatchSource:0}: Error finding container 3d065435931ba0372ae31bb912a5e6a9655979c289b202c89eb5dbc2079f71d9: Status 404 returned error can't find the container with id 3d065435931ba0372ae31bb912a5e6a9655979c289b202c89eb5dbc2079f71d9 Oct 14 15:14:31 crc kubenswrapper[4945]: W1014 15:14:31.257617 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-d1b105ff5b49243fa9d48dcfe74d9d170900382dfbaf0ec64d6149dfabae8985 WatchSource:0}: Error finding container d1b105ff5b49243fa9d48dcfe74d9d170900382dfbaf0ec64d6149dfabae8985: Status 404 returned error can't find the container with id d1b105ff5b49243fa9d48dcfe74d9d170900382dfbaf0ec64d6149dfabae8985 Oct 14 15:14:31 crc kubenswrapper[4945]: W1014 15:14:31.263558 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ec748983f03580be87a59f54141084f82935f60ba34f7c31fa42e9a80a90227b WatchSource:0}: Error finding container ec748983f03580be87a59f54141084f82935f60ba34f7c31fa42e9a80a90227b: Status 404 returned error can't find the container with id ec748983f03580be87a59f54141084f82935f60ba34f7c31fa42e9a80a90227b Oct 14 15:14:31 crc kubenswrapper[4945]: E1014 15:14:31.300443 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.245:6443: connect: connection refused" interval="800ms" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.528034 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.529624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.529681 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.529697 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.529736 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 15:14:31 crc kubenswrapper[4945]: E1014 15:14:31.530393 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.245:6443: connect: connection refused" node="crc" Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.692151 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.766721 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ec748983f03580be87a59f54141084f82935f60ba34f7c31fa42e9a80a90227b"} Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.768507 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d1b105ff5b49243fa9d48dcfe74d9d170900382dfbaf0ec64d6149dfabae8985"} Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.769429 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3d065435931ba0372ae31bb912a5e6a9655979c289b202c89eb5dbc2079f71d9"} Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.770954 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"60092bc1686735e72e267ef3854e7d6725316aae38b943aed170058bb174e6fe"} Oct 14 15:14:31 crc kubenswrapper[4945]: I1014 15:14:31.771998 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2b9270bc97ef89ed5a7cfec7778208e7401791587988a126a96736b95b7266d5"} Oct 14 15:14:31 crc kubenswrapper[4945]: W1014 15:14:31.902884 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:31 crc kubenswrapper[4945]: E1014 15:14:31.903299 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 14 15:14:32 crc kubenswrapper[4945]: E1014 15:14:32.101511 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.245:6443: connect: connection refused" interval="1.6s" Oct 14 15:14:32 crc kubenswrapper[4945]: W1014 15:14:32.112266 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:32 crc kubenswrapper[4945]: E1014 15:14:32.112327 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 14 15:14:32 crc kubenswrapper[4945]: W1014 15:14:32.140624 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:32 crc kubenswrapper[4945]: E1014 15:14:32.140698 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 14 15:14:32 crc kubenswrapper[4945]: W1014 15:14:32.179680 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:32 crc kubenswrapper[4945]: E1014 15:14:32.179782 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.331000 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.332971 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.333039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.333058 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.333099 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 15:14:32 crc kubenswrapper[4945]: E1014 15:14:32.333752 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.245:6443: connect: connection refused" node="crc" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.692220 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.775783 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b" exitCode=0 Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.775896 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.775919 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b"} Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.776825 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.776850 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.776859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.777574 4945 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="63e163a31c424a20cb600a399a42d4dfac6c7f6717ce0060bb2c93cace5f9808" exitCode=0 Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.777645 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"63e163a31c424a20cb600a399a42d4dfac6c7f6717ce0060bb2c93cace5f9808"} Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.777708 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.779159 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.779254 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.779274 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.779284 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.780187 4945 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="6be28c1218c1a6d3076c3cb17faac5f45232bc665d9590e92e2ae3319775f14d" exitCode=0 Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.780230 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.780247 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.780256 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.780255 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"6be28c1218c1a6d3076c3cb17faac5f45232bc665d9590e92e2ae3319775f14d"} Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.780310 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.781524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.781546 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.781556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.782357 4945 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2" exitCode=0 Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.782411 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2"} Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.782480 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.784690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.784712 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.784720 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.788168 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a"} Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.788241 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.788421 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5"} Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.788448 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3"} Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.788458 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33"} Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.789045 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.789110 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:32 crc kubenswrapper[4945]: I1014 15:14:32.789129 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.693072 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:33 crc kubenswrapper[4945]: E1014 15:14:33.702593 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.245:6443: connect: connection refused" interval="3.2s" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.792764 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.795790 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467"} Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.795853 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b"} Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.795886 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e"} Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.796062 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.797512 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.797554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.797569 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.808814 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8"} Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.808901 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb"} Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.808917 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce"} Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.808930 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d"} Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.815130 4945 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a937cba49b54336c332505f45c203d51d827bc244aa8e862a30f88495da9e3a4" exitCode=0 Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.815191 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a937cba49b54336c332505f45c203d51d827bc244aa8e862a30f88495da9e3a4"} Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.815333 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.816118 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.816142 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.816153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.818457 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.818461 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.818440 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3dbbd5b1a7af200f7b8fe75b4235faf670536f59af96ed16e739611b5243ee6f"} Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.819441 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.819473 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.819485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.819454 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.819580 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.819629 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:33 crc kubenswrapper[4945]: W1014 15:14:33.857787 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:33 crc kubenswrapper[4945]: E1014 15:14:33.857897 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.934355 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.936924 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.936962 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.936971 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:33 crc kubenswrapper[4945]: I1014 15:14:33.936993 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 15:14:33 crc kubenswrapper[4945]: E1014 15:14:33.937404 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.245:6443: connect: connection refused" node="crc" Oct 14 15:14:34 crc kubenswrapper[4945]: W1014 15:14:34.205591 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:34 crc kubenswrapper[4945]: E1014 15:14:34.205731 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 14 15:14:34 crc kubenswrapper[4945]: W1014 15:14:34.238989 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 14 15:14:34 crc kubenswrapper[4945]: E1014 15:14:34.239113 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.823819 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98"} Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.823971 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.825146 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.825180 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.825193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.827515 4945 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1cbdeab3be3948d6ca4d4bf7d9b8688b462217b3a9578dd461dc684d7ab685d0" exitCode=0 Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.827553 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1cbdeab3be3948d6ca4d4bf7d9b8688b462217b3a9578dd461dc684d7ab685d0"} Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.827600 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.827629 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.827636 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.827659 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.827682 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.828801 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.828811 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.828824 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.828843 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.828859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.828885 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.828896 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.828906 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.828893 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.829198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.829210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.829219 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:34 crc kubenswrapper[4945]: I1014 15:14:34.939364 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.837921 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d6723f33a9cc957471ffc3bedff560433fa4f63b6b8c282dc56207db2c026f44"} Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.838015 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8b3fec5ff2dbc76d2293f01979ab461f9a2cfa581bdf96af64019df6d6ce7455"} Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.838044 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0bf81d0cfca208a39889da4ed74d6ed868e7a8b215aa8e309add1d23f5ec75c6"} Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.838071 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3603d0b28a1b1f49107575b608c108072f5524627d40a0858b7230355762c78a"} Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.838081 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.838184 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.838257 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.840145 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.840189 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.840240 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.840265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.840313 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.840336 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:35 crc kubenswrapper[4945]: I1014 15:14:35.923691 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:36 crc kubenswrapper[4945]: I1014 15:14:36.793142 4945 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 14 15:14:36 crc kubenswrapper[4945]: I1014 15:14:36.793595 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 14 15:14:36 crc kubenswrapper[4945]: I1014 15:14:36.846685 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"73369395fae4d3939e46a3d49be427526fdf6ff019968997bca7276d3cce4fdb"} Oct 14 15:14:36 crc kubenswrapper[4945]: I1014 15:14:36.846771 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:36 crc kubenswrapper[4945]: I1014 15:14:36.846795 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:36 crc kubenswrapper[4945]: I1014 15:14:36.848156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:36 crc kubenswrapper[4945]: I1014 15:14:36.848193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:36 crc kubenswrapper[4945]: I1014 15:14:36.848203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:36 crc kubenswrapper[4945]: I1014 15:14:36.848273 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:36 crc kubenswrapper[4945]: I1014 15:14:36.848314 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:36 crc kubenswrapper[4945]: I1014 15:14:36.848332 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.138214 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.140244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.140306 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.140321 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.140356 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.278728 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.848806 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.848862 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.850042 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.850089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.850106 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.850534 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.850563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:37 crc kubenswrapper[4945]: I1014 15:14:37.850575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:38 crc kubenswrapper[4945]: I1014 15:14:38.425221 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:38 crc kubenswrapper[4945]: I1014 15:14:38.425512 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:38 crc kubenswrapper[4945]: I1014 15:14:38.426767 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:38 crc kubenswrapper[4945]: I1014 15:14:38.426809 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:38 crc kubenswrapper[4945]: I1014 15:14:38.426825 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:38 crc kubenswrapper[4945]: I1014 15:14:38.851993 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:38 crc kubenswrapper[4945]: I1014 15:14:38.853287 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:38 crc kubenswrapper[4945]: I1014 15:14:38.853347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:38 crc kubenswrapper[4945]: I1014 15:14:38.853361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:39 crc kubenswrapper[4945]: I1014 15:14:39.803457 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:39 crc kubenswrapper[4945]: I1014 15:14:39.803729 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:39 crc kubenswrapper[4945]: I1014 15:14:39.805688 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:39 crc kubenswrapper[4945]: I1014 15:14:39.805772 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:39 crc kubenswrapper[4945]: I1014 15:14:39.805808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:40 crc kubenswrapper[4945]: I1014 15:14:40.782297 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:40 crc kubenswrapper[4945]: I1014 15:14:40.782513 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:40 crc kubenswrapper[4945]: I1014 15:14:40.784097 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:40 crc kubenswrapper[4945]: I1014 15:14:40.784151 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:40 crc kubenswrapper[4945]: I1014 15:14:40.784168 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:40 crc kubenswrapper[4945]: I1014 15:14:40.790634 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:40 crc kubenswrapper[4945]: E1014 15:14:40.830113 4945 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 14 15:14:40 crc kubenswrapper[4945]: I1014 15:14:40.856606 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:40 crc kubenswrapper[4945]: I1014 15:14:40.857403 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:40 crc kubenswrapper[4945]: I1014 15:14:40.857440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:40 crc kubenswrapper[4945]: I1014 15:14:40.857451 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:41 crc kubenswrapper[4945]: I1014 15:14:41.380413 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:41 crc kubenswrapper[4945]: I1014 15:14:41.859203 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:41 crc kubenswrapper[4945]: I1014 15:14:41.860367 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:41 crc kubenswrapper[4945]: I1014 15:14:41.860394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:41 crc kubenswrapper[4945]: I1014 15:14:41.860402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:41 crc kubenswrapper[4945]: I1014 15:14:41.863620 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:42 crc kubenswrapper[4945]: I1014 15:14:42.861939 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:42 crc kubenswrapper[4945]: I1014 15:14:42.862678 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:42 crc kubenswrapper[4945]: I1014 15:14:42.862704 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:42 crc kubenswrapper[4945]: I1014 15:14:42.862712 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:44 crc kubenswrapper[4945]: W1014 15:14:44.562528 4945 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 14 15:14:44 crc kubenswrapper[4945]: I1014 15:14:44.562607 4945 trace.go:236] Trace[1597467030]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 15:14:34.561) (total time: 10001ms): Oct 14 15:14:44 crc kubenswrapper[4945]: Trace[1597467030]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (15:14:44.562) Oct 14 15:14:44 crc kubenswrapper[4945]: Trace[1597467030]: [10.001039373s] [10.001039373s] END Oct 14 15:14:44 crc kubenswrapper[4945]: E1014 15:14:44.562628 4945 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 14 15:14:44 crc kubenswrapper[4945]: I1014 15:14:44.692994 4945 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 14 15:14:44 crc kubenswrapper[4945]: I1014 15:14:44.903755 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51284->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 14 15:14:44 crc kubenswrapper[4945]: I1014 15:14:44.903814 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51284->192.168.126.11:17697: read: connection reset by peer" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.414543 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.414598 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.423262 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.423447 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.424601 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.424649 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.424662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.427342 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.427392 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.477826 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.871747 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.875841 4945 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98" exitCode=255 Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.875928 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98"} Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.876071 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.876071 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.877377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.877401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.877410 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.877378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.877439 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.877455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.878017 4945 scope.go:117] "RemoveContainer" containerID="00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.905857 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.930068 4945 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]log ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]etcd ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/generic-apiserver-start-informers ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/priority-and-fairness-filter ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/start-apiextensions-informers ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/start-apiextensions-controllers ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/crd-informer-synced ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/start-system-namespaces-controller ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 14 15:14:45 crc kubenswrapper[4945]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 14 15:14:45 crc kubenswrapper[4945]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/bootstrap-controller ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/start-kube-aggregator-informers ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/apiservice-registration-controller ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/apiservice-discovery-controller ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]autoregister-completion ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/apiservice-openapi-controller ok Oct 14 15:14:45 crc kubenswrapper[4945]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 14 15:14:45 crc kubenswrapper[4945]: livez check failed Oct 14 15:14:45 crc kubenswrapper[4945]: I1014 15:14:45.930176 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:14:46 crc kubenswrapper[4945]: I1014 15:14:46.793314 4945 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 14 15:14:46 crc kubenswrapper[4945]: I1014 15:14:46.793441 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 14 15:14:46 crc kubenswrapper[4945]: I1014 15:14:46.881016 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 14 15:14:46 crc kubenswrapper[4945]: I1014 15:14:46.883543 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1"} Oct 14 15:14:46 crc kubenswrapper[4945]: I1014 15:14:46.883650 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:46 crc kubenswrapper[4945]: I1014 15:14:46.883744 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:46 crc kubenswrapper[4945]: I1014 15:14:46.886666 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:46 crc kubenswrapper[4945]: I1014 15:14:46.886713 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:46 crc kubenswrapper[4945]: I1014 15:14:46.886717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:46 crc kubenswrapper[4945]: I1014 15:14:46.886744 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:46 crc kubenswrapper[4945]: I1014 15:14:46.887066 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:46 crc kubenswrapper[4945]: I1014 15:14:46.886995 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:48 crc kubenswrapper[4945]: I1014 15:14:48.069011 4945 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.408609 4945 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.420129 4945 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.420647 4945 trace.go:236] Trace[304751376]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 15:14:40.202) (total time: 10218ms): Oct 14 15:14:50 crc kubenswrapper[4945]: Trace[304751376]: ---"Objects listed" error: 10218ms (15:14:50.420) Oct 14 15:14:50 crc kubenswrapper[4945]: Trace[304751376]: [10.218080932s] [10.218080932s] END Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.420703 4945 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.421237 4945 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.421246 4945 trace.go:236] Trace[340781058]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 15:14:38.397) (total time: 12023ms): Oct 14 15:14:50 crc kubenswrapper[4945]: Trace[340781058]: ---"Objects listed" error: 12023ms (15:14:50.421) Oct 14 15:14:50 crc kubenswrapper[4945]: Trace[340781058]: [12.023850128s] [12.023850128s] END Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.421314 4945 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.421445 4945 trace.go:236] Trace[428304173]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 15:14:38.900) (total time: 11521ms): Oct 14 15:14:50 crc kubenswrapper[4945]: Trace[428304173]: ---"Objects listed" error: 11520ms (15:14:50.421) Oct 14 15:14:50 crc kubenswrapper[4945]: Trace[428304173]: [11.521006181s] [11.521006181s] END Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.421467 4945 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.684118 4945 apiserver.go:52] "Watching apiserver" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.688248 4945 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.688549 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.688904 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.689027 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.689188 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.689232 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.689287 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.689321 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.689369 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.689391 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.689450 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.692162 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.692325 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.692545 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.692651 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.692739 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.692847 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.693178 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.693316 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.696168 4945 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.696542 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723185 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723221 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723239 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723255 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723271 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723286 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723303 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723318 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723333 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723349 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723363 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723378 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723504 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723563 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723572 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723663 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723699 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723726 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723747 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724779 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724808 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724827 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724843 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724861 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724892 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724908 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723846 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723860 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723932 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723961 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.723956 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724034 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724066 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724085 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724212 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724246 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724617 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724729 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724903 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.724925 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725060 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725084 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725106 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725125 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725145 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725162 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725190 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725206 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725199 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725230 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725243 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725251 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725274 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725298 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725317 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725337 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725339 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725356 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725378 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725383 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725399 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725420 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725423 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725441 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725467 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725470 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725487 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725507 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725526 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725544 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725565 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725586 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725598 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725607 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725613 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725634 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725657 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725680 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725704 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725718 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725724 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725731 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725746 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725768 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725798 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725820 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725896 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725922 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725945 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725966 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725988 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726014 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726035 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726055 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726076 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726121 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726141 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726161 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726183 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726204 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726225 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726247 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726273 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726295 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726316 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726337 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726358 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726386 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726409 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726430 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726454 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726475 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726498 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726519 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726542 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726567 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726588 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726610 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726632 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726655 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726677 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726699 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726723 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726744 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726765 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726808 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726829 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726851 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726888 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726910 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726934 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726956 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726986 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727007 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727028 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727049 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727068 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727091 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727114 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727135 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727156 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727176 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727198 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727220 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727241 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727289 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727312 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727338 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727360 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727381 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727403 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727483 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727507 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727529 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727552 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727575 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727597 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727655 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727681 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727708 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727731 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727752 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727775 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727799 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727822 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727842 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727864 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728123 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728151 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728175 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728200 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728223 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728248 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728283 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728310 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728334 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728356 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728378 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728400 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725893 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725908 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.725957 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726030 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726107 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726378 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726410 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.730045 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726525 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726621 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726643 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726768 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726842 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726848 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727118 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727371 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727441 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.730206 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727805 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.727811 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728032 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728255 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728456 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728504 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728657 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728752 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.728915 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.729109 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.729153 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.729307 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.729394 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.729592 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.729652 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.729953 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.729985 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.726501 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.730334 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.730347 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.730442 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.730501 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.730630 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.730653 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.730724 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731019 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731083 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731088 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731368 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731496 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731579 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731630 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731666 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731716 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731785 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731809 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731805 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731832 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731859 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731905 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731939 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731963 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.731990 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.732023 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:14:51.231995646 +0000 UTC m=+21.216044104 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732077 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732124 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732159 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732201 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732235 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732267 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732272 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732305 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732337 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732371 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732404 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732431 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732463 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732494 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732525 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732562 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732593 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732624 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732657 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732694 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732727 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732763 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732799 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732832 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732904 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732942 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732975 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733010 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733042 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733076 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733109 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733143 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733173 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733206 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733238 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733301 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733336 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733366 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733437 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733476 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733514 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733549 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733586 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733620 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733658 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733692 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733725 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733790 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733822 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.733990 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734014 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734034 4945 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734052 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734069 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734087 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734103 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734121 4945 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734139 4945 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734158 4945 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734175 4945 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734192 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734209 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734226 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734243 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734261 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734278 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734296 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734313 4945 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734329 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734346 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734365 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734382 4945 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734399 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734419 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734437 4945 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734454 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734471 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734487 4945 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734507 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734528 4945 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734546 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734565 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734582 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734598 4945 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734622 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734639 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734657 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734676 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734692 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734711 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734728 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734744 4945 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734761 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734778 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734794 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734811 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734828 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734846 4945 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.734864 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.735094 4945 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.735117 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.735136 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.735678 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.735727 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.735780 4945 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.735799 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.735848 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.735865 4945 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.735903 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.735924 4945 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.735942 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.743796 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.750655 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.751301 4945 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.768095 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.768451 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.732727 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.739262 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.739363 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.740216 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.740554 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.740717 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.740727 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.741008 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.741033 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.741053 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.741291 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.741944 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.742743 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.742852 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.742945 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.742991 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.743604 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.743889 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.742730 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.744206 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.749197 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.749417 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.750603 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.759236 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.745252 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.760993 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.761004 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.761246 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.761436 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.761747 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.761792 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.764158 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.765596 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.745370 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.766072 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.770140 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.772164 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.772503 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.772931 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.774499 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.774636 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.774812 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.775263 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.775394 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.775469 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:51.275450307 +0000 UTC m=+21.259498685 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.775488 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.775500 4945 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.746099 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.744376 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.767192 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.775531 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.745652 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.767805 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.767839 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.767868 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.767935 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.767989 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.768456 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.768627 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.769120 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.769301 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.744941 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.770086 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.775668 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.775713 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:51.275695313 +0000 UTC m=+21.259743751 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.775841 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.775867 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777210 4945 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777233 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777247 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777273 4945 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777285 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777297 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777310 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777323 4945 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777335 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777347 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777363 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777375 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.777387 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.778233 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.778518 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.778635 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.779320 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.780944 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.787112 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.789026 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.789089 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.789300 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.794298 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.794405 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.794903 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:51.294545029 +0000 UTC m=+21.278593397 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.789653 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.790328 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.789555 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.789580 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.789919 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.790166 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.790378 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.793117 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.793390 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.795710 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.795780 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:50 crc kubenswrapper[4945]: E1014 15:14:50.795927 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:51.295915847 +0000 UTC m=+21.279964215 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.793989 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.796025 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.789101 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.797436 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.798754 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.798775 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.799297 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.799552 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.799620 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.799699 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.800053 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.800364 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.800580 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.800723 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.800910 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.800931 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.801074 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.801291 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.801349 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.801485 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.802695 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.802948 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.802952 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.803194 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.803256 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.803328 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.803470 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.803345 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.803789 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.803913 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.803979 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.803987 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.804001 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.804196 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.804404 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.804439 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.804682 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.805594 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.806721 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.807429 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.811376 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.811411 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.811798 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.812146 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.820424 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.820969 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.821506 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.822648 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.825242 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.827466 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.827505 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.827788 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.832537 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.832984 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.838311 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.838799 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.839377 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.839942 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.840959 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.841398 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.842992 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.843900 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.853177 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.855359 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.857395 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.859157 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.860067 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.860584 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.864702 4945 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.864812 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.871103 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.872254 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.872638 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.876810 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.877275 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.877861 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.877984 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878052 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878084 4945 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878099 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878110 4945 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878122 4945 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878134 4945 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878145 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878156 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878168 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878178 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878189 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878200 4945 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878210 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878222 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878255 4945 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878267 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878278 4945 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878288 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878299 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878310 4945 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878322 4945 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878409 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878434 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878534 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.878641 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879142 4945 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879175 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879188 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879200 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879212 4945 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879223 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879233 4945 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879264 4945 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879275 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879286 4945 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879298 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879309 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879321 4945 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879332 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879345 4945 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879356 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879366 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879378 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879388 4945 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879400 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879413 4945 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879425 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879435 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879447 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879459 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879470 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879481 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879504 4945 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879517 4945 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879528 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879539 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879566 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879598 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879610 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879621 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879633 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879646 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879657 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879681 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879692 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879699 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879704 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879739 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879752 4945 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879765 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879776 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879787 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879798 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879809 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879819 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879888 4945 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879901 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879912 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879923 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879933 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879944 4945 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879954 4945 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880230 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.879965 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880356 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880368 4945 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880379 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880389 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880400 4945 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880428 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880440 4945 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880452 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880463 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880473 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880485 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880497 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880508 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880520 4945 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880533 4945 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880545 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880562 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880573 4945 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880584 4945 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880594 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880606 4945 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880618 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880630 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880641 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880652 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880672 4945 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880683 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880703 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880714 4945 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880725 4945 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880736 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880746 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880760 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880770 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.880781 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.882740 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.885249 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.885729 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.886332 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.887321 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.888262 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.889021 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.889644 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.890220 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.892422 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.892919 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.893773 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.894249 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.894511 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.894840 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.896007 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.896450 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.909525 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.925330 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.928840 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.929420 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.934107 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.935637 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.942946 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.952355 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.961330 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.969915 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.980265 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.988917 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:50 crc kubenswrapper[4945]: I1014 15:14:50.997403 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.011608 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.021299 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 15:14:51 crc kubenswrapper[4945]: W1014 15:14:51.023796 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-a73de2b50d63cf50f6770211dbe294a9449d5bcc561568f319d15d10838e7ed1 WatchSource:0}: Error finding container a73de2b50d63cf50f6770211dbe294a9449d5bcc561568f319d15d10838e7ed1: Status 404 returned error can't find the container with id a73de2b50d63cf50f6770211dbe294a9449d5bcc561568f319d15d10838e7ed1 Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.059936 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 15:14:51 crc kubenswrapper[4945]: W1014 15:14:51.078263 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-4c17de883c9da829eb10bf464682a83d7e71832b80a1792c9181c3598d7790c7 WatchSource:0}: Error finding container 4c17de883c9da829eb10bf464682a83d7e71832b80a1792c9181c3598d7790c7: Status 404 returned error can't find the container with id 4c17de883c9da829eb10bf464682a83d7e71832b80a1792c9181c3598d7790c7 Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.083198 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.283822 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.283927 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.283960 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.284059 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.284116 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:52.284097988 +0000 UTC m=+22.268146356 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.284450 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:14:52.284440048 +0000 UTC m=+22.268488416 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.284526 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.284557 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:52.284548641 +0000 UTC m=+22.268597009 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.384769 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.384842 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.384944 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.384959 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.384972 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.384944 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.385006 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.385013 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:52.38499848 +0000 UTC m=+22.369046848 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.385015 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:51 crc kubenswrapper[4945]: E1014 15:14:51.385090 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:52.385038571 +0000 UTC m=+22.369086939 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.686969 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-mgprj"] Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.687331 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mgprj" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.689101 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.689234 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.689245 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.707702 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.718118 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.728447 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.739207 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.751126 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.767763 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.788558 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g59cr\" (UniqueName: \"kubernetes.io/projected/b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d-kube-api-access-g59cr\") pod \"node-resolver-mgprj\" (UID: \"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\") " pod="openshift-dns/node-resolver-mgprj" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.788602 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d-hosts-file\") pod \"node-resolver-mgprj\" (UID: \"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\") " pod="openshift-dns/node-resolver-mgprj" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.791979 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.812570 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.889298 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d-hosts-file\") pod \"node-resolver-mgprj\" (UID: \"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\") " pod="openshift-dns/node-resolver-mgprj" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.889418 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g59cr\" (UniqueName: \"kubernetes.io/projected/b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d-kube-api-access-g59cr\") pod \"node-resolver-mgprj\" (UID: \"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\") " pod="openshift-dns/node-resolver-mgprj" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.889506 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d-hosts-file\") pod \"node-resolver-mgprj\" (UID: \"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\") " pod="openshift-dns/node-resolver-mgprj" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.900486 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f"} Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.900532 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7"} Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.900542 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a73de2b50d63cf50f6770211dbe294a9449d5bcc561568f319d15d10838e7ed1"} Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.902309 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4c17de883c9da829eb10bf464682a83d7e71832b80a1792c9181c3598d7790c7"} Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.904534 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7"} Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.904603 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d37329e165a42ba3c3d983fc838958412a5d285e125de691a8e341ac7722cbd5"} Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.909774 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g59cr\" (UniqueName: \"kubernetes.io/projected/b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d-kube-api-access-g59cr\") pod \"node-resolver-mgprj\" (UID: \"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\") " pod="openshift-dns/node-resolver-mgprj" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.919925 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.930790 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.938791 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.948857 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.961237 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.974222 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.984540 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:51 crc kubenswrapper[4945]: I1014 15:14:51.993557 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.000981 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mgprj" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.004050 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.020196 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.032601 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.051134 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.073039 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.079798 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-b76xs"] Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.080098 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.084391 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.084450 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.084558 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.084718 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.084720 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-lnbkr"] Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.084962 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.085336 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-tx7tr"] Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.085544 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.085690 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.085551 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7pqtt"] Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.086587 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.087507 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.087940 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.088270 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.088383 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.088460 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.088627 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.088703 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.088845 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.089015 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.089160 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.089467 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.091690 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.091740 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.096129 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.103361 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.123480 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.136320 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.153060 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.169357 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.184454 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.191980 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-multus-socket-dir-parent\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192024 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-slash\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192070 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-cni-bin\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192094 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-multus-cni-dir\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192116 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-systemd\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192140 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-ovn\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192162 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-env-overrides\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192211 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f82b9cd9-7a14-4dd9-bb0e-fde942389666-multus-daemon-config\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192236 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-os-release\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192271 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192368 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-run-k8s-cni-cncf-io\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192425 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-run-netns\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192442 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-log-socket\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192457 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62tc9\" (UniqueName: \"kubernetes.io/projected/6de8ddf6-38db-4394-8660-7e9659bf2bd2-kube-api-access-62tc9\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192473 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blqcc\" (UniqueName: \"kubernetes.io/projected/f82b9cd9-7a14-4dd9-bb0e-fde942389666-kube-api-access-blqcc\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192522 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-hostroot\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192538 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-run-multus-certs\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192553 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-var-lib-openvswitch\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192661 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-run-netns\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192694 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-cni-binary-copy\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192738 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nhpg\" (UniqueName: \"kubernetes.io/projected/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-kube-api-access-8nhpg\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192756 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-cnibin\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192789 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovnkube-config\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192828 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/08049862-1ea7-4efc-b9ed-df61b47d430b-proxy-tls\") pod \"machine-config-daemon-tx7tr\" (UID: \"08049862-1ea7-4efc-b9ed-df61b47d430b\") " pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192845 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-etc-kubernetes\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192867 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-openvswitch\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192913 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192937 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qch55\" (UniqueName: \"kubernetes.io/projected/08049862-1ea7-4efc-b9ed-df61b47d430b-kube-api-access-qch55\") pod \"machine-config-daemon-tx7tr\" (UID: \"08049862-1ea7-4efc-b9ed-df61b47d430b\") " pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192964 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-system-cni-dir\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192982 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-systemd-units\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.192997 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-node-log\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193016 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-system-cni-dir\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193030 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-var-lib-cni-bin\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193044 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-var-lib-cni-multus\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193058 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193075 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-run-ovn-kubernetes\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193092 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-os-release\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193113 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-multus-conf-dir\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193134 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-kubelet\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193182 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-var-lib-kubelet\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193218 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-cnibin\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193253 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-cni-netd\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193275 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovnkube-script-lib\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193308 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/08049862-1ea7-4efc-b9ed-df61b47d430b-mcd-auth-proxy-config\") pod \"machine-config-daemon-tx7tr\" (UID: \"08049862-1ea7-4efc-b9ed-df61b47d430b\") " pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193332 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-etc-openvswitch\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193353 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f82b9cd9-7a14-4dd9-bb0e-fde942389666-cni-binary-copy\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193373 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovn-node-metrics-cert\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.193412 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/08049862-1ea7-4efc-b9ed-df61b47d430b-rootfs\") pod \"machine-config-daemon-tx7tr\" (UID: \"08049862-1ea7-4efc-b9ed-df61b47d430b\") " pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.195473 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.218776 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.230522 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.249983 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.265258 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.293758 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.293850 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-etc-openvswitch\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.293893 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f82b9cd9-7a14-4dd9-bb0e-fde942389666-cni-binary-copy\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.293919 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovn-node-metrics-cert\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.293951 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-etc-openvswitch\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.293961 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:14:54.293934324 +0000 UTC m=+24.277982682 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294040 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/08049862-1ea7-4efc-b9ed-df61b47d430b-rootfs\") pod \"machine-config-daemon-tx7tr\" (UID: \"08049862-1ea7-4efc-b9ed-df61b47d430b\") " pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294059 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-multus-socket-dir-parent\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294074 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-slash\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294083 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/08049862-1ea7-4efc-b9ed-df61b47d430b-rootfs\") pod \"machine-config-daemon-tx7tr\" (UID: \"08049862-1ea7-4efc-b9ed-df61b47d430b\") " pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294089 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-cni-bin\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294122 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-multus-socket-dir-parent\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294106 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-cni-bin\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294151 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-multus-cni-dir\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294170 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-systemd\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294193 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-ovn\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294179 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-slash\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294254 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-systemd\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294208 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-env-overrides\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294297 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f82b9cd9-7a14-4dd9-bb0e-fde942389666-multus-daemon-config\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294318 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-os-release\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294337 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294360 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294366 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-multus-cni-dir\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294378 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-run-k8s-cni-cncf-io\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294393 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-run-netns\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294398 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-ovn\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294408 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-log-socket\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294423 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62tc9\" (UniqueName: \"kubernetes.io/projected/6de8ddf6-38db-4394-8660-7e9659bf2bd2-kube-api-access-62tc9\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294439 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blqcc\" (UniqueName: \"kubernetes.io/projected/f82b9cd9-7a14-4dd9-bb0e-fde942389666-kube-api-access-blqcc\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294480 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-hostroot\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294503 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-run-multus-certs\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294522 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-var-lib-openvswitch\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294543 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-run-netns\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294562 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-cni-binary-copy\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294581 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nhpg\" (UniqueName: \"kubernetes.io/projected/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-kube-api-access-8nhpg\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294599 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-cnibin\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294613 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovnkube-config\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294628 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/08049862-1ea7-4efc-b9ed-df61b47d430b-proxy-tls\") pod \"machine-config-daemon-tx7tr\" (UID: \"08049862-1ea7-4efc-b9ed-df61b47d430b\") " pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294645 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-etc-kubernetes\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294661 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-openvswitch\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294678 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294693 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qch55\" (UniqueName: \"kubernetes.io/projected/08049862-1ea7-4efc-b9ed-df61b47d430b-kube-api-access-qch55\") pod \"machine-config-daemon-tx7tr\" (UID: \"08049862-1ea7-4efc-b9ed-df61b47d430b\") " pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294718 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-run-netns\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294719 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-system-cni-dir\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294742 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-system-cni-dir\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294749 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-systemd-units\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294765 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-node-log\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294782 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-system-cni-dir\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294799 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-var-lib-cni-bin\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294814 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-var-lib-cni-multus\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294830 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294850 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-run-ovn-kubernetes\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294884 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-os-release\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294903 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-multus-conf-dir\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294916 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-kubelet\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294934 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294950 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-var-lib-kubelet\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294964 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-cnibin\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294980 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-cni-netd\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294999 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovnkube-script-lib\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295021 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/08049862-1ea7-4efc-b9ed-df61b47d430b-mcd-auth-proxy-config\") pod \"machine-config-daemon-tx7tr\" (UID: \"08049862-1ea7-4efc-b9ed-df61b47d430b\") " pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295067 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f82b9cd9-7a14-4dd9-bb0e-fde942389666-multus-daemon-config\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295090 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295135 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-system-cni-dir\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.295200 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.295238 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:54.29522592 +0000 UTC m=+24.279274278 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295266 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-os-release\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295297 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-multus-conf-dir\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295320 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-var-lib-cni-bin\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295339 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-var-lib-cni-multus\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295370 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-cni-binary-copy\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295507 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-run-k8s-cni-cncf-io\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295509 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/08049862-1ea7-4efc-b9ed-df61b47d430b-mcd-auth-proxy-config\") pod \"machine-config-daemon-tx7tr\" (UID: \"08049862-1ea7-4efc-b9ed-df61b47d430b\") " pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295539 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-kubelet\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295555 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-run-netns\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.295573 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295582 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-log-socket\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.295597 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:54.295589431 +0000 UTC m=+24.279637799 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295616 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-var-lib-kubelet\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295639 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-cnibin\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295668 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-cni-netd\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295684 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-cnibin\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295783 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295825 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-run-ovn-kubernetes\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295895 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-os-release\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295897 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-hostroot\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295926 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-openvswitch\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295942 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-host-run-multus-certs\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.295976 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-var-lib-openvswitch\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.296001 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-systemd-units\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.296021 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-node-log\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.296050 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f82b9cd9-7a14-4dd9-bb0e-fde942389666-etc-kubernetes\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.296108 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovnkube-script-lib\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.296142 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.294693 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-env-overrides\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.296577 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f82b9cd9-7a14-4dd9-bb0e-fde942389666-cni-binary-copy\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.297799 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovnkube-config\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.299179 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovn-node-metrics-cert\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.310464 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/08049862-1ea7-4efc-b9ed-df61b47d430b-proxy-tls\") pod \"machine-config-daemon-tx7tr\" (UID: \"08049862-1ea7-4efc-b9ed-df61b47d430b\") " pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.323139 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.329232 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blqcc\" (UniqueName: \"kubernetes.io/projected/f82b9cd9-7a14-4dd9-bb0e-fde942389666-kube-api-access-blqcc\") pod \"multus-b76xs\" (UID: \"f82b9cd9-7a14-4dd9-bb0e-fde942389666\") " pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.329278 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62tc9\" (UniqueName: \"kubernetes.io/projected/6de8ddf6-38db-4394-8660-7e9659bf2bd2-kube-api-access-62tc9\") pod \"ovnkube-node-7pqtt\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.329442 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qch55\" (UniqueName: \"kubernetes.io/projected/08049862-1ea7-4efc-b9ed-df61b47d430b-kube-api-access-qch55\") pod \"machine-config-daemon-tx7tr\" (UID: \"08049862-1ea7-4efc-b9ed-df61b47d430b\") " pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.329868 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nhpg\" (UniqueName: \"kubernetes.io/projected/e5ecfd12-913b-4802-99cc-11dfb3d1a0f6-kube-api-access-8nhpg\") pod \"multus-additional-cni-plugins-lnbkr\" (UID: \"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\") " pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.344892 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.357646 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.368584 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.395922 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.395972 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.396238 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.396242 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.396252 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.396260 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.396264 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.396270 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.396312 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:54.396298387 +0000 UTC m=+24.380346755 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.396324 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:54.396319497 +0000 UTC m=+24.380367855 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.397857 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-b76xs" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.406510 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" Oct 14 15:14:52 crc kubenswrapper[4945]: W1014 15:14:52.408606 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf82b9cd9_7a14_4dd9_bb0e_fde942389666.slice/crio-9412110c93c97aca4482fdf2fd7ccfbda687284eb099f13a32aed2bd81073a96 WatchSource:0}: Error finding container 9412110c93c97aca4482fdf2fd7ccfbda687284eb099f13a32aed2bd81073a96: Status 404 returned error can't find the container with id 9412110c93c97aca4482fdf2fd7ccfbda687284eb099f13a32aed2bd81073a96 Oct 14 15:14:52 crc kubenswrapper[4945]: W1014 15:14:52.414944 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5ecfd12_913b_4802_99cc_11dfb3d1a0f6.slice/crio-a98fc0a6326ce92be49ad038710e0798ee972b44daaade32c0333fb78d6f61bc WatchSource:0}: Error finding container a98fc0a6326ce92be49ad038710e0798ee972b44daaade32c0333fb78d6f61bc: Status 404 returned error can't find the container with id a98fc0a6326ce92be49ad038710e0798ee972b44daaade32c0333fb78d6f61bc Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.415525 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.424860 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:52 crc kubenswrapper[4945]: W1014 15:14:52.430278 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08049862_1ea7_4efc_b9ed_df61b47d430b.slice/crio-6a092699f79bfce54aa59f7e973a45b75d79de4f1411922df829adcfab1285eb WatchSource:0}: Error finding container 6a092699f79bfce54aa59f7e973a45b75d79de4f1411922df829adcfab1285eb: Status 404 returned error can't find the container with id 6a092699f79bfce54aa59f7e973a45b75d79de4f1411922df829adcfab1285eb Oct 14 15:14:52 crc kubenswrapper[4945]: W1014 15:14:52.473074 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6de8ddf6_38db_4394_8660_7e9659bf2bd2.slice/crio-aac58667e86956813631c75488951a17c47e7d516188cc5e5ee747d28143da9b WatchSource:0}: Error finding container aac58667e86956813631c75488951a17c47e7d516188cc5e5ee747d28143da9b: Status 404 returned error can't find the container with id aac58667e86956813631c75488951a17c47e7d516188cc5e5ee747d28143da9b Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.761863 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.761922 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.761916 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.762036 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.762095 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:14:52 crc kubenswrapper[4945]: E1014 15:14:52.762144 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.766681 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.767766 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.768488 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.769301 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.770821 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.771557 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.772693 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.773376 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.908122 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883" exitCode=0 Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.908187 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883"} Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.908222 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerStarted","Data":"aac58667e86956813631c75488951a17c47e7d516188cc5e5ee747d28143da9b"} Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.909829 4945 generic.go:334] "Generic (PLEG): container finished" podID="e5ecfd12-913b-4802-99cc-11dfb3d1a0f6" containerID="30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b" exitCode=0 Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.909927 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" event={"ID":"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6","Type":"ContainerDied","Data":"30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b"} Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.909993 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" event={"ID":"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6","Type":"ContainerStarted","Data":"a98fc0a6326ce92be49ad038710e0798ee972b44daaade32c0333fb78d6f61bc"} Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.911198 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mgprj" event={"ID":"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d","Type":"ContainerStarted","Data":"2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106"} Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.911239 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mgprj" event={"ID":"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d","Type":"ContainerStarted","Data":"9721af7904829f522687a8a03feab0256b47862127250d0b3e4f9a6a1bc4adeb"} Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.912532 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b76xs" event={"ID":"f82b9cd9-7a14-4dd9-bb0e-fde942389666","Type":"ContainerStarted","Data":"42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289"} Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.912556 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b76xs" event={"ID":"f82b9cd9-7a14-4dd9-bb0e-fde942389666","Type":"ContainerStarted","Data":"9412110c93c97aca4482fdf2fd7ccfbda687284eb099f13a32aed2bd81073a96"} Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.914156 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerStarted","Data":"cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f"} Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.914181 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerStarted","Data":"e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd"} Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.914193 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerStarted","Data":"6a092699f79bfce54aa59f7e973a45b75d79de4f1411922df829adcfab1285eb"} Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.924861 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.942661 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.955467 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.968924 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:52 crc kubenswrapper[4945]: I1014 15:14:52.998218 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:52Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.010555 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.024937 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.034536 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.047085 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.058366 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.069326 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.082304 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.095127 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.108590 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.134296 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.151202 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.164211 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.177034 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.194388 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.223204 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.253221 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.272845 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.292578 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.313903 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.796976 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.801417 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.809795 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.815809 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.828649 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.842475 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.856434 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.867887 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.884458 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.898140 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.910320 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.918618 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10"} Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.921924 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerStarted","Data":"251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b"} Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.921946 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerStarted","Data":"3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d"} Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.921956 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerStarted","Data":"8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59"} Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.921965 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerStarted","Data":"85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563"} Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.921973 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerStarted","Data":"7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04"} Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.921981 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerStarted","Data":"e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488"} Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.923614 4945 generic.go:334] "Generic (PLEG): container finished" podID="e5ecfd12-913b-4802-99cc-11dfb3d1a0f6" containerID="66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb" exitCode=0 Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.923992 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" event={"ID":"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6","Type":"ContainerDied","Data":"66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb"} Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.928335 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.940152 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.948972 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.966497 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.973777 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-zfc4b"] Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.974110 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zfc4b" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.975686 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.976099 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.976595 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.977749 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.979075 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:53 crc kubenswrapper[4945]: I1014 15:14:53.990937 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:53Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.001837 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.010926 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.027620 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.044956 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.060168 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.073696 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.085645 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.098818 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.112947 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6snbc\" (UniqueName: \"kubernetes.io/projected/7648f858-37c9-4f14-9d0f-5e3b354f4e47-kube-api-access-6snbc\") pod \"node-ca-zfc4b\" (UID: \"7648f858-37c9-4f14-9d0f-5e3b354f4e47\") " pod="openshift-image-registry/node-ca-zfc4b" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.113059 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7648f858-37c9-4f14-9d0f-5e3b354f4e47-host\") pod \"node-ca-zfc4b\" (UID: \"7648f858-37c9-4f14-9d0f-5e3b354f4e47\") " pod="openshift-image-registry/node-ca-zfc4b" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.113082 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7648f858-37c9-4f14-9d0f-5e3b354f4e47-serviceca\") pod \"node-ca-zfc4b\" (UID: \"7648f858-37c9-4f14-9d0f-5e3b354f4e47\") " pod="openshift-image-registry/node-ca-zfc4b" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.120360 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.133243 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.144255 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.161845 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.179358 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.195050 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.205278 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.214530 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6snbc\" (UniqueName: \"kubernetes.io/projected/7648f858-37c9-4f14-9d0f-5e3b354f4e47-kube-api-access-6snbc\") pod \"node-ca-zfc4b\" (UID: \"7648f858-37c9-4f14-9d0f-5e3b354f4e47\") " pod="openshift-image-registry/node-ca-zfc4b" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.214661 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7648f858-37c9-4f14-9d0f-5e3b354f4e47-host\") pod \"node-ca-zfc4b\" (UID: \"7648f858-37c9-4f14-9d0f-5e3b354f4e47\") " pod="openshift-image-registry/node-ca-zfc4b" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.214699 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7648f858-37c9-4f14-9d0f-5e3b354f4e47-serviceca\") pod \"node-ca-zfc4b\" (UID: \"7648f858-37c9-4f14-9d0f-5e3b354f4e47\") " pod="openshift-image-registry/node-ca-zfc4b" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.214830 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7648f858-37c9-4f14-9d0f-5e3b354f4e47-host\") pod \"node-ca-zfc4b\" (UID: \"7648f858-37c9-4f14-9d0f-5e3b354f4e47\") " pod="openshift-image-registry/node-ca-zfc4b" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.216495 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7648f858-37c9-4f14-9d0f-5e3b354f4e47-serviceca\") pod \"node-ca-zfc4b\" (UID: \"7648f858-37c9-4f14-9d0f-5e3b354f4e47\") " pod="openshift-image-registry/node-ca-zfc4b" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.219862 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.231905 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6snbc\" (UniqueName: \"kubernetes.io/projected/7648f858-37c9-4f14-9d0f-5e3b354f4e47-kube-api-access-6snbc\") pod \"node-ca-zfc4b\" (UID: \"7648f858-37c9-4f14-9d0f-5e3b354f4e47\") " pod="openshift-image-registry/node-ca-zfc4b" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.234630 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.246502 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.257764 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.266776 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.277809 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.291531 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.303648 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.313809 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.315113 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.315295 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:14:58.315280143 +0000 UTC m=+28.299328511 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.315342 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.315384 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.315445 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.315613 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:58.315592351 +0000 UTC m=+28.299640729 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.315461 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.315665 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:58.315656993 +0000 UTC m=+28.299705371 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.332964 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.352494 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zfc4b" Oct 14 15:14:54 crc kubenswrapper[4945]: W1014 15:14:54.369495 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7648f858_37c9_4f14_9d0f_5e3b354f4e47.slice/crio-af25a19eb31e7810b3ba6c4f34f4d532076f926b9aeb293cadf56cde75656a36 WatchSource:0}: Error finding container af25a19eb31e7810b3ba6c4f34f4d532076f926b9aeb293cadf56cde75656a36: Status 404 returned error can't find the container with id af25a19eb31e7810b3ba6c4f34f4d532076f926b9aeb293cadf56cde75656a36 Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.416364 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.417071 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.416586 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.417288 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.417299 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.417348 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:58.417331866 +0000 UTC m=+28.401380234 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.417263 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.417658 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.417673 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.417702 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 15:14:58.417692816 +0000 UTC m=+28.401741184 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.762058 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.762106 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.762054 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.762202 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.762309 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:14:54 crc kubenswrapper[4945]: E1014 15:14:54.762392 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.929500 4945 generic.go:334] "Generic (PLEG): container finished" podID="e5ecfd12-913b-4802-99cc-11dfb3d1a0f6" containerID="dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046" exitCode=0 Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.929553 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" event={"ID":"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6","Type":"ContainerDied","Data":"dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046"} Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.930954 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zfc4b" event={"ID":"7648f858-37c9-4f14-9d0f-5e3b354f4e47","Type":"ContainerStarted","Data":"5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac"} Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.930991 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zfc4b" event={"ID":"7648f858-37c9-4f14-9d0f-5e3b354f4e47","Type":"ContainerStarted","Data":"af25a19eb31e7810b3ba6c4f34f4d532076f926b9aeb293cadf56cde75656a36"} Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.944476 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.968315 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.985057 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:54 crc kubenswrapper[4945]: I1014 15:14:54.995182 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.004533 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.016776 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.029962 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.043387 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.056633 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.072456 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.097113 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.120115 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.134723 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.162488 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.173141 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.181733 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.204821 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.217014 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.230104 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.240427 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.252291 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.263426 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.274558 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.285403 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.316613 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.356515 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.399014 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.438384 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.937241 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerStarted","Data":"228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0"} Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.939662 4945 generic.go:334] "Generic (PLEG): container finished" podID="e5ecfd12-913b-4802-99cc-11dfb3d1a0f6" containerID="7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15" exitCode=0 Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.939689 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" event={"ID":"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6","Type":"ContainerDied","Data":"7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15"} Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.959979 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.981791 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:55 crc kubenswrapper[4945]: I1014 15:14:55.995850 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:55Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.015861 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.031018 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.042507 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.057707 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.072310 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.083398 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.102028 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.114186 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.125111 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.134740 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.151111 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.762499 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.762594 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:56 crc kubenswrapper[4945]: E1014 15:14:56.762628 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:14:56 crc kubenswrapper[4945]: E1014 15:14:56.762723 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.762791 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:14:56 crc kubenswrapper[4945]: E1014 15:14:56.762852 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.820287 4945 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.822161 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.822196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.822207 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.822310 4945 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.828473 4945 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.828678 4945 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.829471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.829489 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.829496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.829509 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.829517 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:56Z","lastTransitionTime":"2025-10-14T15:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:56 crc kubenswrapper[4945]: E1014 15:14:56.842927 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.847231 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.847265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.847274 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.847286 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.847294 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:56Z","lastTransitionTime":"2025-10-14T15:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:56 crc kubenswrapper[4945]: E1014 15:14:56.860655 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.864279 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.864309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.864321 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.864370 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.864382 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:56Z","lastTransitionTime":"2025-10-14T15:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:56 crc kubenswrapper[4945]: E1014 15:14:56.878817 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.882982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.883118 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.883132 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.883147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.883156 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:56Z","lastTransitionTime":"2025-10-14T15:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:56 crc kubenswrapper[4945]: E1014 15:14:56.896705 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.900093 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.900141 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.900153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.900169 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.900180 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:56Z","lastTransitionTime":"2025-10-14T15:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:56 crc kubenswrapper[4945]: E1014 15:14:56.912527 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: E1014 15:14:56.912896 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.914379 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.914505 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.914604 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.914690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.914755 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:56Z","lastTransitionTime":"2025-10-14T15:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.945800 4945 generic.go:334] "Generic (PLEG): container finished" podID="e5ecfd12-913b-4802-99cc-11dfb3d1a0f6" containerID="e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80" exitCode=0 Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.945893 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" event={"ID":"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6","Type":"ContainerDied","Data":"e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80"} Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.959735 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.979614 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:56 crc kubenswrapper[4945]: I1014 15:14:56.989450 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.001881 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.011845 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.018345 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.018390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.018402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.018419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.018432 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:57Z","lastTransitionTime":"2025-10-14T15:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.024289 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.053543 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.097354 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.109082 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.120365 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.120401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.120425 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.120436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.120451 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.120462 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:57Z","lastTransitionTime":"2025-10-14T15:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.131903 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.144153 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.153463 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.171763 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.222928 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.222965 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.222977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.223001 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.223015 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:57Z","lastTransitionTime":"2025-10-14T15:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.325149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.325186 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.325195 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.325208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.325217 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:57Z","lastTransitionTime":"2025-10-14T15:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.428988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.429071 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.429092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.429118 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.429137 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:57Z","lastTransitionTime":"2025-10-14T15:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.531501 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.531571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.531592 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.531660 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.531684 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:57Z","lastTransitionTime":"2025-10-14T15:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.634654 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.634725 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.634743 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.634770 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.634787 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:57Z","lastTransitionTime":"2025-10-14T15:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.737987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.738040 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.738057 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.738083 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.738100 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:57Z","lastTransitionTime":"2025-10-14T15:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.841096 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.841156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.841181 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.841210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.841234 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:57Z","lastTransitionTime":"2025-10-14T15:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.944517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.944561 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.944570 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.944586 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.944595 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:57Z","lastTransitionTime":"2025-10-14T15:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.955127 4945 generic.go:334] "Generic (PLEG): container finished" podID="e5ecfd12-913b-4802-99cc-11dfb3d1a0f6" containerID="06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107" exitCode=0 Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.955185 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" event={"ID":"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6","Type":"ContainerDied","Data":"06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107"} Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.977157 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:57 crc kubenswrapper[4945]: I1014 15:14:57.990489 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.001448 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:57Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.019570 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.036425 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.047155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.047184 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.047191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.047205 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.047214 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:58Z","lastTransitionTime":"2025-10-14T15:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.050330 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.063835 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.076747 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.096085 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.111929 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.122444 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.135447 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.147687 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.149153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.149173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.149181 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.149193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.149202 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:58Z","lastTransitionTime":"2025-10-14T15:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.157231 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.251483 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.251503 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.251511 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.251522 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.251531 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:58Z","lastTransitionTime":"2025-10-14T15:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.350176 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.350349 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.350397 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:15:06.350366898 +0000 UTC m=+36.334415276 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.350507 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.350574 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.350584 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:06.350560643 +0000 UTC m=+36.334609051 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.350618 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:06.350607394 +0000 UTC m=+36.334655772 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.350502 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.354578 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.354629 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.354649 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.355018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.355048 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:58Z","lastTransitionTime":"2025-10-14T15:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.451385 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.451472 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.451646 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.451665 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.451686 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.451701 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.451704 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.451723 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.451787 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:06.451759853 +0000 UTC m=+36.435808261 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.451832 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:06.451819354 +0000 UTC m=+36.435867812 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.458090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.458125 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.458135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.458167 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.458182 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:58Z","lastTransitionTime":"2025-10-14T15:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.561364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.561409 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.561421 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.561440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.561452 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:58Z","lastTransitionTime":"2025-10-14T15:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.664547 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.664597 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.664609 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.664627 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.664639 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:58Z","lastTransitionTime":"2025-10-14T15:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.761960 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.761989 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.762110 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.762199 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.762294 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:14:58 crc kubenswrapper[4945]: E1014 15:14:58.762524 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.767485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.767551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.767570 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.767593 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.767625 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:58Z","lastTransitionTime":"2025-10-14T15:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.870610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.870665 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.870680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.870699 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.870712 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:58Z","lastTransitionTime":"2025-10-14T15:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.964851 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerStarted","Data":"cf67567c04110801ffc8d222dd992d92aedfb3c8ab58af8f542ce341b20b8ce8"} Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.965245 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.970479 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" event={"ID":"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6","Type":"ContainerStarted","Data":"34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028"} Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.972301 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.972339 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.972350 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.972363 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.972374 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:58Z","lastTransitionTime":"2025-10-14T15:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.979416 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.991174 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:58 crc kubenswrapper[4945]: I1014 15:14:58.994243 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.006742 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf67567c04110801ffc8d222dd992d92aedfb3c8ab58af8f542ce341b20b8ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.019599 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.033763 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.043284 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.053075 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.071128 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.074862 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.074910 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.074937 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.074955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.074965 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:59Z","lastTransitionTime":"2025-10-14T15:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.087529 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.098991 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.110795 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.122701 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.134501 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.146808 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.163181 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.176520 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.176865 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.176921 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.176932 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.176948 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.176959 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:59Z","lastTransitionTime":"2025-10-14T15:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.187765 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.204775 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf67567c04110801ffc8d222dd992d92aedfb3c8ab58af8f542ce341b20b8ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.217400 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.230772 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.243993 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.261583 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.275169 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.278669 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.278727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.278737 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.278753 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.278763 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:59Z","lastTransitionTime":"2025-10-14T15:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.287099 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.297561 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.306283 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.318816 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.330130 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:14:59Z is after 2025-08-24T17:21:41Z" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.381020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.381068 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.381079 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.381095 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.381108 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:59Z","lastTransitionTime":"2025-10-14T15:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.483273 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.483317 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.483328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.483344 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.483355 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:59Z","lastTransitionTime":"2025-10-14T15:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.585650 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.585695 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.585705 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.585721 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.585729 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:59Z","lastTransitionTime":"2025-10-14T15:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.687235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.687269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.687279 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.687292 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.687302 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:59Z","lastTransitionTime":"2025-10-14T15:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.790560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.790610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.790623 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.790641 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.790652 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:59Z","lastTransitionTime":"2025-10-14T15:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.893220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.893269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.893304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.893324 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.893338 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:59Z","lastTransitionTime":"2025-10-14T15:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.975327 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.976112 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.996560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.996619 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.996636 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.996662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:14:59 crc kubenswrapper[4945]: I1014 15:14:59.996679 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:14:59Z","lastTransitionTime":"2025-10-14T15:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.011095 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.027856 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.042822 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.056934 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.075506 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf67567c04110801ffc8d222dd992d92aedfb3c8ab58af8f542ce341b20b8ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.089916 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.098921 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.098987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.098996 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.099011 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.099023 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:00Z","lastTransitionTime":"2025-10-14T15:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.100869 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.116366 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.143770 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.154576 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.168430 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.183232 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.198197 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.201969 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.202006 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.202017 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.202037 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.202055 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:00Z","lastTransitionTime":"2025-10-14T15:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.216777 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.229989 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.304649 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.304727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.304752 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.304779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.304808 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:00Z","lastTransitionTime":"2025-10-14T15:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.407494 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.407554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.407572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.407596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.407611 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:00Z","lastTransitionTime":"2025-10-14T15:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.510166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.510199 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.510208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.510223 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.510234 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:00Z","lastTransitionTime":"2025-10-14T15:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.612297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.612352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.612366 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.612386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.612398 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:00Z","lastTransitionTime":"2025-10-14T15:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.714890 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.714937 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.714951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.714971 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.714983 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:00Z","lastTransitionTime":"2025-10-14T15:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.762334 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:00 crc kubenswrapper[4945]: E1014 15:15:00.762479 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.762555 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.762706 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:00 crc kubenswrapper[4945]: E1014 15:15:00.762947 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:00 crc kubenswrapper[4945]: E1014 15:15:00.763082 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.782096 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.796345 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.807398 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.816901 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.816943 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.816955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.816971 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.816982 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:00Z","lastTransitionTime":"2025-10-14T15:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.817426 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.826134 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.842562 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.856353 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.867930 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.880507 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.893072 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.906169 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.919305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.919350 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.919361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.919378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.919389 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:00Z","lastTransitionTime":"2025-10-14T15:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.919438 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.930192 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.952615 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf67567c04110801ffc8d222dd992d92aedfb3c8ab58af8f542ce341b20b8ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:00 crc kubenswrapper[4945]: I1014 15:15:00.978555 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.021866 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.021925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.021937 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.021950 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.021958 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:01Z","lastTransitionTime":"2025-10-14T15:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.123980 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.124241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.124250 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.124265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.124275 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:01Z","lastTransitionTime":"2025-10-14T15:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.227023 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.227082 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.227100 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.227122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.227136 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:01Z","lastTransitionTime":"2025-10-14T15:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.329341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.329590 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.329680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.329815 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.329909 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:01Z","lastTransitionTime":"2025-10-14T15:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.431777 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.431840 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.431850 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.431864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.431904 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:01Z","lastTransitionTime":"2025-10-14T15:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.534781 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.534834 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.534853 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.534911 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.534929 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:01Z","lastTransitionTime":"2025-10-14T15:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.638208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.638491 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.638609 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.638762 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.638912 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:01Z","lastTransitionTime":"2025-10-14T15:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.741309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.741392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.741417 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.741449 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.741471 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:01Z","lastTransitionTime":"2025-10-14T15:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.843725 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.843754 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.843762 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.843774 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.843783 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:01Z","lastTransitionTime":"2025-10-14T15:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.946004 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.946039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.946052 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.946066 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.946076 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:01Z","lastTransitionTime":"2025-10-14T15:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.983431 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/0.log" Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.987251 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="cf67567c04110801ffc8d222dd992d92aedfb3c8ab58af8f542ce341b20b8ce8" exitCode=1 Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.987319 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"cf67567c04110801ffc8d222dd992d92aedfb3c8ab58af8f542ce341b20b8ce8"} Oct 14 15:15:01 crc kubenswrapper[4945]: I1014 15:15:01.988162 4945 scope.go:117] "RemoveContainer" containerID="cf67567c04110801ffc8d222dd992d92aedfb3c8ab58af8f542ce341b20b8ce8" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.001527 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:01Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.017619 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.033100 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.046381 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.048417 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.048442 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.048451 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.048463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.048472 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:02Z","lastTransitionTime":"2025-10-14T15:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.059857 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.070847 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.084228 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.101803 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.117732 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.130427 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.148590 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf67567c04110801ffc8d222dd992d92aedfb3c8ab58af8f542ce341b20b8ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf67567c04110801ffc8d222dd992d92aedfb3c8ab58af8f542ce341b20b8ce8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:01Z\\\",\\\"message\\\":\\\" 6240 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 15:15:01.114154 6240 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 15:15:01.114168 6240 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1014 15:15:01.114194 6240 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 15:15:01.114202 6240 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 15:15:01.114221 6240 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 15:15:01.114254 6240 factory.go:656] Stopping watch factory\\\\nI1014 15:15:01.114260 6240 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 15:15:01.114272 6240 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 15:15:01.114274 6240 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 15:15:01.114219 6240 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 15:15:01.114270 6240 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.150988 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.151026 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.151036 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.151051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.151061 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:02Z","lastTransitionTime":"2025-10-14T15:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.168589 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.184025 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.196924 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.253454 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.253517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.253533 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.253555 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.253571 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:02Z","lastTransitionTime":"2025-10-14T15:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.355736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.355797 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.355814 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.355839 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.355857 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:02Z","lastTransitionTime":"2025-10-14T15:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.457552 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.457583 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.457594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.457608 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.457619 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:02Z","lastTransitionTime":"2025-10-14T15:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.559971 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.560004 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.560016 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.560031 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.560043 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:02Z","lastTransitionTime":"2025-10-14T15:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.662040 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.662093 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.662105 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.662121 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.662133 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:02Z","lastTransitionTime":"2025-10-14T15:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.762342 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.762420 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.762415 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:02 crc kubenswrapper[4945]: E1014 15:15:02.762492 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:02 crc kubenswrapper[4945]: E1014 15:15:02.762595 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:02 crc kubenswrapper[4945]: E1014 15:15:02.762671 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.764144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.764178 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.764192 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.764208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.764221 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:02Z","lastTransitionTime":"2025-10-14T15:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.867020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.867067 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.867077 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.867092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.867103 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:02Z","lastTransitionTime":"2025-10-14T15:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.986548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.986590 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.986603 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.986621 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.986632 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:02Z","lastTransitionTime":"2025-10-14T15:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.991064 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/1.log" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.991666 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/0.log" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.998633 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8" exitCode=1 Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.998668 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8"} Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.998710 4945 scope.go:117] "RemoveContainer" containerID="cf67567c04110801ffc8d222dd992d92aedfb3c8ab58af8f542ce341b20b8ce8" Oct 14 15:15:02 crc kubenswrapper[4945]: I1014 15:15:02.999467 4945 scope.go:117] "RemoveContainer" containerID="3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8" Oct 14 15:15:02 crc kubenswrapper[4945]: E1014 15:15:02.999633 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.045441 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.063554 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.082475 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf67567c04110801ffc8d222dd992d92aedfb3c8ab58af8f542ce341b20b8ce8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:01Z\\\",\\\"message\\\":\\\" 6240 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 15:15:01.114154 6240 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 15:15:01.114168 6240 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1014 15:15:01.114194 6240 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 15:15:01.114202 6240 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 15:15:01.114221 6240 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 15:15:01.114254 6240 factory.go:656] Stopping watch factory\\\\nI1014 15:15:01.114260 6240 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 15:15:01.114272 6240 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 15:15:01.114274 6240 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 15:15:01.114219 6240 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 15:15:01.114270 6240 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:02Z\\\",\\\"message\\\":\\\"calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z]\\\\nI1014 15:15:02.926934 6379 services_controller.go:434] Service openshift-machine-api/cluster-autoscaler-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{cluster-autoscaler-operator openshift-machine-api b8e0040a-0eca-4299-ac4a-f26a24879998 4394 0 2025-02-23 05:12:30 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:cluster-autoscaler-operator] map[exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:cluster-autoscaler-operator-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007ae224b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.088356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.088389 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.088399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.088413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.088423 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:03Z","lastTransitionTime":"2025-10-14T15:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.094630 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.107248 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.117540 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.126757 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.143411 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.162240 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.177769 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.190025 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.190587 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.190621 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.190632 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.190648 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.190659 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:03Z","lastTransitionTime":"2025-10-14T15:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.203037 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.217281 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.232129 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:03Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.293150 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.293190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.293202 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.293219 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.293230 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:03Z","lastTransitionTime":"2025-10-14T15:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.395148 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.395204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.395216 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.395236 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.395249 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:03Z","lastTransitionTime":"2025-10-14T15:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.497165 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.497200 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.497210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.497224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.497232 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:03Z","lastTransitionTime":"2025-10-14T15:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.599394 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.599660 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.599728 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.599861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.599942 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:03Z","lastTransitionTime":"2025-10-14T15:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.702177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.702212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.702222 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.702236 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.702246 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:03Z","lastTransitionTime":"2025-10-14T15:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.805671 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.806022 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.806121 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.806238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.806328 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:03Z","lastTransitionTime":"2025-10-14T15:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.912144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.912214 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.912237 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.912264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:03 crc kubenswrapper[4945]: I1014 15:15:03.912285 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:03Z","lastTransitionTime":"2025-10-14T15:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.004389 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/1.log" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.007815 4945 scope.go:117] "RemoveContainer" containerID="3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8" Oct 14 15:15:04 crc kubenswrapper[4945]: E1014 15:15:04.007969 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.014091 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.014116 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.014134 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.014152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.014163 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:04Z","lastTransitionTime":"2025-10-14T15:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.021020 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.037289 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.055315 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.071092 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.088267 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.104072 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.117230 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.117269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.117281 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.117301 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.117314 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:04Z","lastTransitionTime":"2025-10-14T15:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.121858 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.138362 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.152313 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.164808 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.189645 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:02Z\\\",\\\"message\\\":\\\"calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z]\\\\nI1014 15:15:02.926934 6379 services_controller.go:434] Service openshift-machine-api/cluster-autoscaler-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{cluster-autoscaler-operator openshift-machine-api b8e0040a-0eca-4299-ac4a-f26a24879998 4394 0 2025-02-23 05:12:30 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:cluster-autoscaler-operator] map[exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:cluster-autoscaler-operator-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007ae224b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.209925 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.219778 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.219833 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.219851 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.219907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.219924 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:04Z","lastTransitionTime":"2025-10-14T15:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.233979 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.249037 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.322063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.322125 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.322143 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.322167 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.322185 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:04Z","lastTransitionTime":"2025-10-14T15:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.425515 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.425584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.425607 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.425641 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.425664 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:04Z","lastTransitionTime":"2025-10-14T15:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.529203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.529283 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.529310 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.529342 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.529364 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:04Z","lastTransitionTime":"2025-10-14T15:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.632662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.632760 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.632787 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.632816 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.632841 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:04Z","lastTransitionTime":"2025-10-14T15:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.713385 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.736239 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.736281 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.736291 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.736308 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.736319 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:04Z","lastTransitionTime":"2025-10-14T15:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.749115 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.759971 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.762094 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.762141 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.762146 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:04 crc kubenswrapper[4945]: E1014 15:15:04.762303 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:04 crc kubenswrapper[4945]: E1014 15:15:04.762388 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:04 crc kubenswrapper[4945]: E1014 15:15:04.762461 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.778563 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.792504 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.793266 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd"] Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.793684 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.796170 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.796234 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.816670 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.827735 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.838504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.838542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.838556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.838572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.838585 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:04Z","lastTransitionTime":"2025-10-14T15:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.840497 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.852559 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.866337 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.880995 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.892978 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.911312 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.919861 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a336c34-8669-4fbe-8f51-c339403782b6-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bmsrd\" (UID: \"5a336c34-8669-4fbe-8f51-c339403782b6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.919925 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a336c34-8669-4fbe-8f51-c339403782b6-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bmsrd\" (UID: \"5a336c34-8669-4fbe-8f51-c339403782b6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.919979 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a336c34-8669-4fbe-8f51-c339403782b6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bmsrd\" (UID: \"5a336c34-8669-4fbe-8f51-c339403782b6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.920022 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk2nw\" (UniqueName: \"kubernetes.io/projected/5a336c34-8669-4fbe-8f51-c339403782b6-kube-api-access-lk2nw\") pod \"ovnkube-control-plane-749d76644c-bmsrd\" (UID: \"5a336c34-8669-4fbe-8f51-c339403782b6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.925256 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.940732 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.940774 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.940787 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.940805 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.940818 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:04Z","lastTransitionTime":"2025-10-14T15:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.947698 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:02Z\\\",\\\"message\\\":\\\"calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z]\\\\nI1014 15:15:02.926934 6379 services_controller.go:434] Service openshift-machine-api/cluster-autoscaler-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{cluster-autoscaler-operator openshift-machine-api b8e0040a-0eca-4299-ac4a-f26a24879998 4394 0 2025-02-23 05:12:30 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:cluster-autoscaler-operator] map[exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:cluster-autoscaler-operator-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007ae224b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.960435 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:04 crc kubenswrapper[4945]: I1014 15:15:04.983819 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:04Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.005111 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.020927 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a336c34-8669-4fbe-8f51-c339403782b6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bmsrd\" (UID: \"5a336c34-8669-4fbe-8f51-c339403782b6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.021010 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk2nw\" (UniqueName: \"kubernetes.io/projected/5a336c34-8669-4fbe-8f51-c339403782b6-kube-api-access-lk2nw\") pod \"ovnkube-control-plane-749d76644c-bmsrd\" (UID: \"5a336c34-8669-4fbe-8f51-c339403782b6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.021070 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a336c34-8669-4fbe-8f51-c339403782b6-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bmsrd\" (UID: \"5a336c34-8669-4fbe-8f51-c339403782b6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.021106 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a336c34-8669-4fbe-8f51-c339403782b6-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bmsrd\" (UID: \"5a336c34-8669-4fbe-8f51-c339403782b6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.022430 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a336c34-8669-4fbe-8f51-c339403782b6-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bmsrd\" (UID: \"5a336c34-8669-4fbe-8f51-c339403782b6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.022738 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a336c34-8669-4fbe-8f51-c339403782b6-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bmsrd\" (UID: \"5a336c34-8669-4fbe-8f51-c339403782b6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.028096 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a336c34-8669-4fbe-8f51-c339403782b6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bmsrd\" (UID: \"5a336c34-8669-4fbe-8f51-c339403782b6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.028689 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.043328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.043370 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.043387 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.043413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.043431 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:05Z","lastTransitionTime":"2025-10-14T15:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.043983 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.051607 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk2nw\" (UniqueName: \"kubernetes.io/projected/5a336c34-8669-4fbe-8f51-c339403782b6-kube-api-access-lk2nw\") pod \"ovnkube-control-plane-749d76644c-bmsrd\" (UID: \"5a336c34-8669-4fbe-8f51-c339403782b6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.060536 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.072979 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.085385 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.099783 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.109612 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.111326 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: W1014 15:15:05.130647 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a336c34_8669_4fbe_8f51_c339403782b6.slice/crio-7eff02e983643920fbc111289b672a21e22e54174630bdc0eda96f09adb1aa42 WatchSource:0}: Error finding container 7eff02e983643920fbc111289b672a21e22e54174630bdc0eda96f09adb1aa42: Status 404 returned error can't find the container with id 7eff02e983643920fbc111289b672a21e22e54174630bdc0eda96f09adb1aa42 Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.134031 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:02Z\\\",\\\"message\\\":\\\"calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z]\\\\nI1014 15:15:02.926934 6379 services_controller.go:434] Service openshift-machine-api/cluster-autoscaler-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{cluster-autoscaler-operator openshift-machine-api b8e0040a-0eca-4299-ac4a-f26a24879998 4394 0 2025-02-23 05:12:30 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:cluster-autoscaler-operator] map[exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:cluster-autoscaler-operator-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007ae224b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.148411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.148443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.148451 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.148483 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.148525 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:05Z","lastTransitionTime":"2025-10-14T15:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.151147 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.164919 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.176410 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.186543 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.250750 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.250804 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.250819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.250836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.250848 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:05Z","lastTransitionTime":"2025-10-14T15:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.352820 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.352855 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.352864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.352891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.352902 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:05Z","lastTransitionTime":"2025-10-14T15:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.455229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.455273 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.455284 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.455302 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.455314 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:05Z","lastTransitionTime":"2025-10-14T15:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.558107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.558157 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.558169 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.558187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.558199 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:05Z","lastTransitionTime":"2025-10-14T15:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.660801 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.660912 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.660950 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.660980 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.661001 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:05Z","lastTransitionTime":"2025-10-14T15:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.763098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.763141 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.763152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.763172 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.763185 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:05Z","lastTransitionTime":"2025-10-14T15:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.866009 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.866050 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.866062 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.866084 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.866096 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:05Z","lastTransitionTime":"2025-10-14T15:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.890426 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-ftg5h"] Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.891702 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:05 crc kubenswrapper[4945]: E1014 15:15:05.891813 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.913662 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.933785 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.954715 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.965435 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.968144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.968175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.968187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.968204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.968216 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:05Z","lastTransitionTime":"2025-10-14T15:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.977154 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:05 crc kubenswrapper[4945]: I1014 15:15:05.989244 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.000471 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:05Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.015347 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" event={"ID":"5a336c34-8669-4fbe-8f51-c339403782b6","Type":"ContainerStarted","Data":"e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59"} Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.015397 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" event={"ID":"5a336c34-8669-4fbe-8f51-c339403782b6","Type":"ContainerStarted","Data":"62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0"} Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.015412 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" event={"ID":"5a336c34-8669-4fbe-8f51-c339403782b6","Type":"ContainerStarted","Data":"7eff02e983643920fbc111289b672a21e22e54174630bdc0eda96f09adb1aa42"} Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.017806 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.030954 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvc8w\" (UniqueName: \"kubernetes.io/projected/be17bc84-909e-4150-8ee7-9a7d23173ba0-kube-api-access-jvc8w\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.031010 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.033925 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.046499 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.066807 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:02Z\\\",\\\"message\\\":\\\"calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z]\\\\nI1014 15:15:02.926934 6379 services_controller.go:434] Service openshift-machine-api/cluster-autoscaler-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{cluster-autoscaler-operator openshift-machine-api b8e0040a-0eca-4299-ac4a-f26a24879998 4394 0 2025-02-23 05:12:30 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:cluster-autoscaler-operator] map[exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:cluster-autoscaler-operator-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007ae224b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.070453 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.070503 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.070520 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.070540 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.070554 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:06Z","lastTransitionTime":"2025-10-14T15:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.084919 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.098634 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.109663 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.120110 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.132153 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvc8w\" (UniqueName: \"kubernetes.io/projected/be17bc84-909e-4150-8ee7-9a7d23173ba0-kube-api-access-jvc8w\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.132208 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.132391 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.132469 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs podName:be17bc84-909e-4150-8ee7-9a7d23173ba0 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:06.6324474 +0000 UTC m=+36.616495788 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs") pod "network-metrics-daemon-ftg5h" (UID: "be17bc84-909e-4150-8ee7-9a7d23173ba0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.135238 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.148346 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.149393 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvc8w\" (UniqueName: \"kubernetes.io/projected/be17bc84-909e-4150-8ee7-9a7d23173ba0-kube-api-access-jvc8w\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.160239 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.171059 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.172156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.172190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.172206 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.172228 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.172239 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:06Z","lastTransitionTime":"2025-10-14T15:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.181348 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.196796 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:02Z\\\",\\\"message\\\":\\\"calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z]\\\\nI1014 15:15:02.926934 6379 services_controller.go:434] Service openshift-machine-api/cluster-autoscaler-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{cluster-autoscaler-operator openshift-machine-api b8e0040a-0eca-4299-ac4a-f26a24879998 4394 0 2025-02-23 05:12:30 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:cluster-autoscaler-operator] map[exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:cluster-autoscaler-operator-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007ae224b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.208175 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.219865 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.229711 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.238730 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.247582 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.258125 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.270505 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.274098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.274134 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.274144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.274158 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.274169 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:06Z","lastTransitionTime":"2025-10-14T15:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.282365 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.292240 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.302133 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.310579 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:06Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.376271 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.376324 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.376338 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.376361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.376378 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:06Z","lastTransitionTime":"2025-10-14T15:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.434825 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.435047 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:15:22.43500508 +0000 UTC m=+52.419053488 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.435197 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.435309 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.435411 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.435470 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.435506 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:22.435486473 +0000 UTC m=+52.419534891 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.435547 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:22.435527465 +0000 UTC m=+52.419575873 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.480044 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.480106 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.480122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.480164 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.480183 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:06Z","lastTransitionTime":"2025-10-14T15:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.536822 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.536918 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.537045 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.537065 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.537078 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.537094 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.537133 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.537149 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.537134 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:22.537117395 +0000 UTC m=+52.521165763 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.537276 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:22.537240679 +0000 UTC m=+52.521289097 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.583223 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.583293 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.583318 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.583347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.583368 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:06Z","lastTransitionTime":"2025-10-14T15:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.637943 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.638096 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.638494 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs podName:be17bc84-909e-4150-8ee7-9a7d23173ba0 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:07.638468429 +0000 UTC m=+37.622516837 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs") pod "network-metrics-daemon-ftg5h" (UID: "be17bc84-909e-4150-8ee7-9a7d23173ba0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.690807 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.690864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.690919 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.690952 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.690975 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:06Z","lastTransitionTime":"2025-10-14T15:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.762416 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.762463 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.762584 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.762946 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.763089 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:06 crc kubenswrapper[4945]: E1014 15:15:06.763249 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.793991 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.794043 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.794054 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.794075 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.794087 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:06Z","lastTransitionTime":"2025-10-14T15:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.896726 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.896772 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.896787 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.896805 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:06 crc kubenswrapper[4945]: I1014 15:15:06.896819 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:06Z","lastTransitionTime":"2025-10-14T15:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:06.999975 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.000349 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.000569 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.000781 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.001027 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.068827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.068907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.068930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.068953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.068969 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: E1014 15:15:07.092929 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:07Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.098964 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.099052 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.099076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.099106 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.099128 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: E1014 15:15:07.118591 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:07Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.123977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.124035 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.124057 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.124082 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.124100 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: E1014 15:15:07.142635 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:07Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.147517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.147564 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.147582 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.147607 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.147623 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: E1014 15:15:07.165963 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:07Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.170061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.170107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.170126 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.170149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.170165 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: E1014 15:15:07.183181 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:07Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:07 crc kubenswrapper[4945]: E1014 15:15:07.183467 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.185837 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.185894 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.185903 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.185918 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.185928 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.289268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.289347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.289370 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.289399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.289421 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.391642 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.391709 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.391733 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.391765 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.391789 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.494371 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.494416 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.494427 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.494442 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.494454 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.598485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.598797 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.598934 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.599078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.599205 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.648116 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:07 crc kubenswrapper[4945]: E1014 15:15:07.648257 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:07 crc kubenswrapper[4945]: E1014 15:15:07.648313 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs podName:be17bc84-909e-4150-8ee7-9a7d23173ba0 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:09.648297364 +0000 UTC m=+39.632345732 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs") pod "network-metrics-daemon-ftg5h" (UID: "be17bc84-909e-4150-8ee7-9a7d23173ba0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.701823 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.701860 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.701887 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.701902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.701911 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.762859 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:07 crc kubenswrapper[4945]: E1014 15:15:07.763186 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.804376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.804425 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.804435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.804453 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.804464 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.907339 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.907407 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.907426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.907454 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:07 crc kubenswrapper[4945]: I1014 15:15:07.907473 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:07Z","lastTransitionTime":"2025-10-14T15:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.010237 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.010294 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.010312 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.010341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.010356 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:08Z","lastTransitionTime":"2025-10-14T15:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.118827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.118869 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.118898 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.118916 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.118928 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:08Z","lastTransitionTime":"2025-10-14T15:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.221533 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.221589 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.221600 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.221619 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.221631 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:08Z","lastTransitionTime":"2025-10-14T15:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.324784 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.324853 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.324902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.324936 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.324960 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:08Z","lastTransitionTime":"2025-10-14T15:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.428046 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.428089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.428100 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.428114 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.428125 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:08Z","lastTransitionTime":"2025-10-14T15:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.531222 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.531294 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.531319 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.531352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.531375 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:08Z","lastTransitionTime":"2025-10-14T15:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.633637 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.633678 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.633693 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.633714 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.633730 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:08Z","lastTransitionTime":"2025-10-14T15:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.736783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.736816 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.736827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.736842 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.736853 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:08Z","lastTransitionTime":"2025-10-14T15:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.762769 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:08 crc kubenswrapper[4945]: E1014 15:15:08.762922 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.762987 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.762785 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:08 crc kubenswrapper[4945]: E1014 15:15:08.763162 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:08 crc kubenswrapper[4945]: E1014 15:15:08.763306 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.840602 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.840756 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.840786 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.840859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.840915 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:08Z","lastTransitionTime":"2025-10-14T15:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.945153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.945236 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.945261 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.945297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:08 crc kubenswrapper[4945]: I1014 15:15:08.945318 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:08Z","lastTransitionTime":"2025-10-14T15:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.048861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.048931 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.048943 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.048962 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.048976 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:09Z","lastTransitionTime":"2025-10-14T15:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.151694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.151742 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.151759 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.151782 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.151797 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:09Z","lastTransitionTime":"2025-10-14T15:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.254319 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.254606 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.254745 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.254903 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.254992 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:09Z","lastTransitionTime":"2025-10-14T15:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.357776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.358159 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.358285 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.358403 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.358511 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:09Z","lastTransitionTime":"2025-10-14T15:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.461410 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.461695 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.461805 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.461927 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.462021 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:09Z","lastTransitionTime":"2025-10-14T15:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.565106 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.565180 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.565201 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.565226 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.565247 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:09Z","lastTransitionTime":"2025-10-14T15:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.667711 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.667770 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.667786 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.667803 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.667815 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:09Z","lastTransitionTime":"2025-10-14T15:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.671368 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:09 crc kubenswrapper[4945]: E1014 15:15:09.671603 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:09 crc kubenswrapper[4945]: E1014 15:15:09.671674 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs podName:be17bc84-909e-4150-8ee7-9a7d23173ba0 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:13.671654789 +0000 UTC m=+43.655703167 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs") pod "network-metrics-daemon-ftg5h" (UID: "be17bc84-909e-4150-8ee7-9a7d23173ba0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.762239 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:09 crc kubenswrapper[4945]: E1014 15:15:09.762434 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.769813 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.769847 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.769864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.769899 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.769908 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:09Z","lastTransitionTime":"2025-10-14T15:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.872291 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.872333 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.872347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.872366 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.872378 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:09Z","lastTransitionTime":"2025-10-14T15:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.975047 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.975078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.975086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.975099 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:09 crc kubenswrapper[4945]: I1014 15:15:09.975109 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:09Z","lastTransitionTime":"2025-10-14T15:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.077791 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.077868 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.077921 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.077953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.077977 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:10Z","lastTransitionTime":"2025-10-14T15:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.180796 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.181494 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.181637 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.181797 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.181938 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:10Z","lastTransitionTime":"2025-10-14T15:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.284552 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.284592 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.284604 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.284620 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.284632 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:10Z","lastTransitionTime":"2025-10-14T15:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.387282 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.387336 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.387354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.387381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.387398 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:10Z","lastTransitionTime":"2025-10-14T15:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.490767 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.490851 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.490904 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.490931 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.490960 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:10Z","lastTransitionTime":"2025-10-14T15:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.593478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.593543 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.593561 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.593585 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.593607 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:10Z","lastTransitionTime":"2025-10-14T15:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.696460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.696510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.696527 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.696551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.696569 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:10Z","lastTransitionTime":"2025-10-14T15:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.762437 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:10 crc kubenswrapper[4945]: E1014 15:15:10.762613 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.762749 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:10 crc kubenswrapper[4945]: E1014 15:15:10.762940 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.763143 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:10 crc kubenswrapper[4945]: E1014 15:15:10.763261 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.786122 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.799345 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.799531 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.799666 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.800058 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.800306 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:10Z","lastTransitionTime":"2025-10-14T15:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.807665 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.826515 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.843655 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.864358 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:02Z\\\",\\\"message\\\":\\\"calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z]\\\\nI1014 15:15:02.926934 6379 services_controller.go:434] Service openshift-machine-api/cluster-autoscaler-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{cluster-autoscaler-operator openshift-machine-api b8e0040a-0eca-4299-ac4a-f26a24879998 4394 0 2025-02-23 05:12:30 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:cluster-autoscaler-operator] map[exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:cluster-autoscaler-operator-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007ae224b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.884661 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.903128 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.903191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.903208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.903230 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.903245 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:10Z","lastTransitionTime":"2025-10-14T15:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.907010 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.920698 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.935919 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.951145 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.964090 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.979564 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:10 crc kubenswrapper[4945]: I1014 15:15:10.994199 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:10Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.005154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.005203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.005214 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.005234 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.005246 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:11Z","lastTransitionTime":"2025-10-14T15:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.005804 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:11Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.018393 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:11Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.028921 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:11Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.109063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.109163 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.109183 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.109209 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.109227 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:11Z","lastTransitionTime":"2025-10-14T15:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.212188 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.212293 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.212313 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.212341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.212358 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:11Z","lastTransitionTime":"2025-10-14T15:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.315181 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.315226 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.315234 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.315246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.315255 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:11Z","lastTransitionTime":"2025-10-14T15:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.419480 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.419545 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.419559 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.419581 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.419592 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:11Z","lastTransitionTime":"2025-10-14T15:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.522514 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.522550 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.522560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.522575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.522584 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:11Z","lastTransitionTime":"2025-10-14T15:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.625043 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.625117 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.625142 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.625172 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.625194 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:11Z","lastTransitionTime":"2025-10-14T15:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.727959 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.728024 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.728049 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.728080 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.728105 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:11Z","lastTransitionTime":"2025-10-14T15:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.762474 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:11 crc kubenswrapper[4945]: E1014 15:15:11.762610 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.829830 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.829865 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.829906 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.829928 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.829942 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:11Z","lastTransitionTime":"2025-10-14T15:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.932579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.932811 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.932944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.933118 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:11 crc kubenswrapper[4945]: I1014 15:15:11.933239 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:11Z","lastTransitionTime":"2025-10-14T15:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.036952 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.037002 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.037016 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.037034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.037047 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:12Z","lastTransitionTime":"2025-10-14T15:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.139702 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.139736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.139744 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.139761 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.139770 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:12Z","lastTransitionTime":"2025-10-14T15:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.242943 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.242989 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.243008 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.243032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.243050 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:12Z","lastTransitionTime":"2025-10-14T15:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.346148 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.346196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.346204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.346219 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.346227 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:12Z","lastTransitionTime":"2025-10-14T15:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.427443 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.429001 4945 scope.go:117] "RemoveContainer" containerID="3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8" Oct 14 15:15:12 crc kubenswrapper[4945]: E1014 15:15:12.429532 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.448516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.448550 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.448562 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.448576 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.448585 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:12Z","lastTransitionTime":"2025-10-14T15:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.551098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.551159 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.551175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.551198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.551214 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:12Z","lastTransitionTime":"2025-10-14T15:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.653769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.653809 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.653819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.653838 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.653850 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:12Z","lastTransitionTime":"2025-10-14T15:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.756001 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.756264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.756326 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.756385 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.756460 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:12Z","lastTransitionTime":"2025-10-14T15:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.762501 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:12 crc kubenswrapper[4945]: E1014 15:15:12.762726 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.763167 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:12 crc kubenswrapper[4945]: E1014 15:15:12.763478 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.763617 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:12 crc kubenswrapper[4945]: E1014 15:15:12.763747 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.859424 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.859693 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.859813 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.859955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.860047 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:12Z","lastTransitionTime":"2025-10-14T15:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.963345 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.963401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.963414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.963434 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:12 crc kubenswrapper[4945]: I1014 15:15:12.963450 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:12Z","lastTransitionTime":"2025-10-14T15:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.065797 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.066185 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.066537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.066825 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.067170 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:13Z","lastTransitionTime":"2025-10-14T15:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.170777 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.170826 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.170839 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.170858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.170890 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:13Z","lastTransitionTime":"2025-10-14T15:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.273775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.273818 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.273828 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.273845 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.273858 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:13Z","lastTransitionTime":"2025-10-14T15:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.376122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.376157 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.376166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.376180 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.376191 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:13Z","lastTransitionTime":"2025-10-14T15:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.478576 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.478824 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.478917 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.479008 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.479068 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:13Z","lastTransitionTime":"2025-10-14T15:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.582075 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.582112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.582121 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.582136 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.582146 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:13Z","lastTransitionTime":"2025-10-14T15:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.684543 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.684602 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.684624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.684649 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.684666 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:13Z","lastTransitionTime":"2025-10-14T15:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.716369 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:13 crc kubenswrapper[4945]: E1014 15:15:13.716534 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:13 crc kubenswrapper[4945]: E1014 15:15:13.716639 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs podName:be17bc84-909e-4150-8ee7-9a7d23173ba0 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:21.716618899 +0000 UTC m=+51.700667267 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs") pod "network-metrics-daemon-ftg5h" (UID: "be17bc84-909e-4150-8ee7-9a7d23173ba0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.762610 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:13 crc kubenswrapper[4945]: E1014 15:15:13.762835 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.787490 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.787563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.787586 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.787614 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.787638 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:13Z","lastTransitionTime":"2025-10-14T15:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.890557 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.890628 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.890646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.890670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.890698 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:13Z","lastTransitionTime":"2025-10-14T15:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.993358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.993411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.993427 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.993452 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:13 crc kubenswrapper[4945]: I1014 15:15:13.993469 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:13Z","lastTransitionTime":"2025-10-14T15:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.096315 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.096352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.096365 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.096387 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.096398 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:14Z","lastTransitionTime":"2025-10-14T15:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.198662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.199030 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.199160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.199257 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.199321 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:14Z","lastTransitionTime":"2025-10-14T15:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.303004 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.303072 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.303090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.303116 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.303137 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:14Z","lastTransitionTime":"2025-10-14T15:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.406063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.406113 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.406125 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.406145 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.406157 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:14Z","lastTransitionTime":"2025-10-14T15:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.509274 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.509360 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.509379 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.509408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.509424 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:14Z","lastTransitionTime":"2025-10-14T15:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.612843 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.612907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.612917 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.612935 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.612947 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:14Z","lastTransitionTime":"2025-10-14T15:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.716060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.716168 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.716196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.716228 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.716251 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:14Z","lastTransitionTime":"2025-10-14T15:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.762376 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.762383 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.762574 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:14 crc kubenswrapper[4945]: E1014 15:15:14.762806 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:14 crc kubenswrapper[4945]: E1014 15:15:14.762989 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:14 crc kubenswrapper[4945]: E1014 15:15:14.763128 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.819288 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.819352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.819373 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.819397 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.819414 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:14Z","lastTransitionTime":"2025-10-14T15:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.923323 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.923393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.923416 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.923544 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:14 crc kubenswrapper[4945]: I1014 15:15:14.923568 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:14Z","lastTransitionTime":"2025-10-14T15:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.026370 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.026428 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.026440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.026463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.026478 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:15Z","lastTransitionTime":"2025-10-14T15:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.129837 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.129915 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.129932 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.129953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.129967 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:15Z","lastTransitionTime":"2025-10-14T15:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.233303 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.233364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.233378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.233397 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.233409 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:15Z","lastTransitionTime":"2025-10-14T15:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.336512 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.337446 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.337691 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.337929 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.338136 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:15Z","lastTransitionTime":"2025-10-14T15:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.441440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.441507 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.441526 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.441551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.441566 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:15Z","lastTransitionTime":"2025-10-14T15:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.544308 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.544378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.544395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.544428 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.544447 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:15Z","lastTransitionTime":"2025-10-14T15:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.646859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.646962 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.646980 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.647008 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.647070 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:15Z","lastTransitionTime":"2025-10-14T15:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.750740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.750807 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.750822 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.750844 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.750859 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:15Z","lastTransitionTime":"2025-10-14T15:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.762039 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:15 crc kubenswrapper[4945]: E1014 15:15:15.762190 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.853919 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.853986 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.854000 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.854029 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.854046 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:15Z","lastTransitionTime":"2025-10-14T15:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.957268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.957318 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.957331 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.957350 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:15 crc kubenswrapper[4945]: I1014 15:15:15.957361 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:15Z","lastTransitionTime":"2025-10-14T15:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.059909 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.059952 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.059966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.059986 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.059998 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:16Z","lastTransitionTime":"2025-10-14T15:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.162715 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.162775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.162789 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.162812 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.162826 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:16Z","lastTransitionTime":"2025-10-14T15:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.266065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.266109 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.266120 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.266169 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.266184 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:16Z","lastTransitionTime":"2025-10-14T15:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.369915 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.370086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.370147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.370189 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.370215 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:16Z","lastTransitionTime":"2025-10-14T15:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.474002 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.474054 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.474064 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.474079 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.474088 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:16Z","lastTransitionTime":"2025-10-14T15:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.577228 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.577493 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.577647 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.577769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.578094 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:16Z","lastTransitionTime":"2025-10-14T15:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.680964 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.681019 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.681032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.681052 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.681067 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:16Z","lastTransitionTime":"2025-10-14T15:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.762786 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.762912 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.762921 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:16 crc kubenswrapper[4945]: E1014 15:15:16.763034 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:16 crc kubenswrapper[4945]: E1014 15:15:16.763336 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:16 crc kubenswrapper[4945]: E1014 15:15:16.763348 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.783834 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.783875 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.783901 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.783916 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.783928 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:16Z","lastTransitionTime":"2025-10-14T15:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.885912 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.885953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.885962 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.885977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.885986 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:16Z","lastTransitionTime":"2025-10-14T15:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.988830 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.988877 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.988906 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.988923 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:16 crc kubenswrapper[4945]: I1014 15:15:16.988937 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:16Z","lastTransitionTime":"2025-10-14T15:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.091932 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.091973 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.091984 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.092003 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.092017 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.193666 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.193951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.194060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.194162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.194243 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.296690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.296751 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.296763 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.296782 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.296793 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.399985 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.401262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.401340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.401444 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.401541 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.504900 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.505269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.505370 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.505506 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.505587 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.507372 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.507439 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.507456 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.507478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.507494 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: E1014 15:15:17.526141 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:17Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.532089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.532435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.532514 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.532621 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.532702 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: E1014 15:15:17.547927 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:17Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.552829 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.552879 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.552891 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.552933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.552954 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: E1014 15:15:17.567191 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:17Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.572293 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.572344 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.572354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.572378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.572393 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: E1014 15:15:17.585974 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:17Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.590461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.590511 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.590530 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.590553 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.590566 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: E1014 15:15:17.605314 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:17Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:17 crc kubenswrapper[4945]: E1014 15:15:17.605438 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.607717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.607789 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.607803 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.607829 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.607845 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.711048 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.711117 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.711132 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.711153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.711164 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.762930 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:17 crc kubenswrapper[4945]: E1014 15:15:17.763122 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.814307 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.814355 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.814364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.814385 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.814397 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.917433 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.917487 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.917532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.917551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:17 crc kubenswrapper[4945]: I1014 15:15:17.917565 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:17Z","lastTransitionTime":"2025-10-14T15:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.020409 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.020711 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.020808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.020949 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.021097 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:18Z","lastTransitionTime":"2025-10-14T15:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.125153 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.125227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.125242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.125269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.125286 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:18Z","lastTransitionTime":"2025-10-14T15:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.227744 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.227780 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.227790 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.227808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.227818 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:18Z","lastTransitionTime":"2025-10-14T15:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.330698 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.330792 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.330807 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.330828 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.330843 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:18Z","lastTransitionTime":"2025-10-14T15:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.433659 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.433721 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.433734 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.433755 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.433767 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:18Z","lastTransitionTime":"2025-10-14T15:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.537405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.537448 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.537458 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.537477 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.537488 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:18Z","lastTransitionTime":"2025-10-14T15:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.640597 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.640847 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.640990 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.641082 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.641146 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:18Z","lastTransitionTime":"2025-10-14T15:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.743305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.743343 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.743354 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.743369 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.743379 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:18Z","lastTransitionTime":"2025-10-14T15:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.762038 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.762069 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:18 crc kubenswrapper[4945]: E1014 15:15:18.762161 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.762233 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:18 crc kubenswrapper[4945]: E1014 15:15:18.762332 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:18 crc kubenswrapper[4945]: E1014 15:15:18.762377 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.845924 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.846152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.846220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.846287 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.846348 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:18Z","lastTransitionTime":"2025-10-14T15:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.948715 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.948806 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.948816 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.948833 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:18 crc kubenswrapper[4945]: I1014 15:15:18.948844 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:18Z","lastTransitionTime":"2025-10-14T15:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.051366 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.051471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.051483 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.051497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.051506 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:19Z","lastTransitionTime":"2025-10-14T15:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.153476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.153524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.153539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.153559 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.153574 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:19Z","lastTransitionTime":"2025-10-14T15:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.256237 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.256299 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.256316 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.256340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.256361 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:19Z","lastTransitionTime":"2025-10-14T15:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.359148 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.359228 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.359246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.359282 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.359300 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:19Z","lastTransitionTime":"2025-10-14T15:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.461838 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.461901 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.461914 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.461933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.461944 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:19Z","lastTransitionTime":"2025-10-14T15:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.564312 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.564363 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.564378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.564398 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.564408 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:19Z","lastTransitionTime":"2025-10-14T15:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.666811 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.666860 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.666880 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.666923 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.666943 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:19Z","lastTransitionTime":"2025-10-14T15:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.762323 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:19 crc kubenswrapper[4945]: E1014 15:15:19.762476 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.770455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.770520 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.770537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.770560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.770579 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:19Z","lastTransitionTime":"2025-10-14T15:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.873065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.873114 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.873127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.873149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.873164 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:19Z","lastTransitionTime":"2025-10-14T15:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.975789 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.975857 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.975868 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.975908 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:19 crc kubenswrapper[4945]: I1014 15:15:19.975921 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:19Z","lastTransitionTime":"2025-10-14T15:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.079406 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.079455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.079469 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.079489 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.079502 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:20Z","lastTransitionTime":"2025-10-14T15:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.182315 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.182347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.182356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.182370 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.182380 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:20Z","lastTransitionTime":"2025-10-14T15:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.284784 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.284828 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.284840 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.284854 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.284866 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:20Z","lastTransitionTime":"2025-10-14T15:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.387501 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.387555 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.387564 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.387577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.387588 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:20Z","lastTransitionTime":"2025-10-14T15:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.490306 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.491576 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.491758 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.492000 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.492248 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:20Z","lastTransitionTime":"2025-10-14T15:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.596041 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.596391 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.596551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.596779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.597219 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:20Z","lastTransitionTime":"2025-10-14T15:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.699777 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.699928 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.699955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.699981 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.699997 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:20Z","lastTransitionTime":"2025-10-14T15:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.762597 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.762720 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:20 crc kubenswrapper[4945]: E1014 15:15:20.763332 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:20 crc kubenswrapper[4945]: E1014 15:15:20.763459 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.762778 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:20 crc kubenswrapper[4945]: E1014 15:15:20.764090 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.779864 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.796925 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.802029 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.802076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.802088 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.802107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.802119 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:20Z","lastTransitionTime":"2025-10-14T15:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.812773 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.825085 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.855432 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:02Z\\\",\\\"message\\\":\\\"calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z]\\\\nI1014 15:15:02.926934 6379 services_controller.go:434] Service openshift-machine-api/cluster-autoscaler-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{cluster-autoscaler-operator openshift-machine-api b8e0040a-0eca-4299-ac4a-f26a24879998 4394 0 2025-02-23 05:12:30 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:cluster-autoscaler-operator] map[exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:cluster-autoscaler-operator-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007ae224b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.871805 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.893173 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.904510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.904577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.904600 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.904631 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.904656 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:20Z","lastTransitionTime":"2025-10-14T15:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.908063 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.920979 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.931986 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.946696 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.959639 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.973862 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:20 crc kubenswrapper[4945]: I1014 15:15:20.987593 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:20Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.002230 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:21Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.006352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.006396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.006408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.006423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.006432 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:21Z","lastTransitionTime":"2025-10-14T15:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.012900 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:21Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.109218 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.109283 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.109302 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.109327 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.109344 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:21Z","lastTransitionTime":"2025-10-14T15:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.213188 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.213236 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.213250 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.213270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.213283 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:21Z","lastTransitionTime":"2025-10-14T15:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.316677 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.316766 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.316855 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.316921 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.316984 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:21Z","lastTransitionTime":"2025-10-14T15:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.419074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.419109 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.419117 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.419130 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.419139 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:21Z","lastTransitionTime":"2025-10-14T15:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.521365 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.522012 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.522119 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.522235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.522521 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:21Z","lastTransitionTime":"2025-10-14T15:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.625575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.625624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.625640 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.625663 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.625676 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:21Z","lastTransitionTime":"2025-10-14T15:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.728521 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.728558 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.728571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.728588 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.728599 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:21Z","lastTransitionTime":"2025-10-14T15:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.762698 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:21 crc kubenswrapper[4945]: E1014 15:15:21.763343 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.812163 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:21 crc kubenswrapper[4945]: E1014 15:15:21.812339 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:21 crc kubenswrapper[4945]: E1014 15:15:21.812397 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs podName:be17bc84-909e-4150-8ee7-9a7d23173ba0 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:37.812379721 +0000 UTC m=+67.796428089 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs") pod "network-metrics-daemon-ftg5h" (UID: "be17bc84-909e-4150-8ee7-9a7d23173ba0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.831569 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.831630 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.831647 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.831668 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.831685 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:21Z","lastTransitionTime":"2025-10-14T15:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.933377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.933442 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.933465 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.933493 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:21 crc kubenswrapper[4945]: I1014 15:15:21.933516 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:21Z","lastTransitionTime":"2025-10-14T15:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.036753 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.036817 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.036827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.036908 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.036919 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:22Z","lastTransitionTime":"2025-10-14T15:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.139649 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.139688 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.139736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.139750 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.139761 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:22Z","lastTransitionTime":"2025-10-14T15:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.243149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.243209 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.243226 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.243249 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.243264 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:22Z","lastTransitionTime":"2025-10-14T15:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.345863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.345910 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.345921 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.345933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.345942 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:22Z","lastTransitionTime":"2025-10-14T15:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.448476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.448925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.449098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.449220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.449368 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:22Z","lastTransitionTime":"2025-10-14T15:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.519810 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.520049 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.520094 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.520215 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.520214 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.520263 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:54.520248213 +0000 UTC m=+84.504296581 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.520341 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:54.520306995 +0000 UTC m=+84.504355353 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.520541 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:15:54.520530821 +0000 UTC m=+84.504579189 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.552713 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.552757 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.552775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.552797 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.552817 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:22Z","lastTransitionTime":"2025-10-14T15:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.621543 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.621599 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.621721 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.621741 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.621753 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.621805 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:54.621790353 +0000 UTC m=+84.605838721 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.622136 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.622231 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.622252 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.622357 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 15:15:54.622328508 +0000 UTC m=+84.606377046 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.655765 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.655812 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.655824 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.655842 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.655854 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:22Z","lastTransitionTime":"2025-10-14T15:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.758495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.758551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.758568 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.758594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.758611 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:22Z","lastTransitionTime":"2025-10-14T15:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.762786 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.762838 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.762806 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.762968 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.763206 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:22 crc kubenswrapper[4945]: E1014 15:15:22.763270 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.861460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.861514 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.861532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.861555 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.861573 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:22Z","lastTransitionTime":"2025-10-14T15:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.964967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.965277 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.965442 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.965601 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:22 crc kubenswrapper[4945]: I1014 15:15:22.965733 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:22Z","lastTransitionTime":"2025-10-14T15:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.070118 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.070171 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.070187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.070210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.070227 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:23Z","lastTransitionTime":"2025-10-14T15:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.172013 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.172051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.172062 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.172077 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.172087 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:23Z","lastTransitionTime":"2025-10-14T15:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.275051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.275535 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.275866 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.276362 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.276722 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:23Z","lastTransitionTime":"2025-10-14T15:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.379986 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.380051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.380061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.380078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.380089 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:23Z","lastTransitionTime":"2025-10-14T15:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.483107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.483159 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.483173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.483192 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.483207 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:23Z","lastTransitionTime":"2025-10-14T15:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.586564 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.586633 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.586656 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.586689 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.586710 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:23Z","lastTransitionTime":"2025-10-14T15:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.689691 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.689792 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.689825 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.689856 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.689918 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:23Z","lastTransitionTime":"2025-10-14T15:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.762583 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:23 crc kubenswrapper[4945]: E1014 15:15:23.762739 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.792497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.792537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.792548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.792567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.792582 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:23Z","lastTransitionTime":"2025-10-14T15:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.894989 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.895025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.895033 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.895046 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.895055 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:23Z","lastTransitionTime":"2025-10-14T15:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.997234 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.997476 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.997550 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.997613 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:23 crc kubenswrapper[4945]: I1014 15:15:23.997671 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:23Z","lastTransitionTime":"2025-10-14T15:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.099912 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.099972 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.099989 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.100010 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.100028 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:24Z","lastTransitionTime":"2025-10-14T15:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.202205 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.202252 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.202268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.202291 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.202310 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:24Z","lastTransitionTime":"2025-10-14T15:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.305102 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.305171 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.305192 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.305218 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.305238 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:24Z","lastTransitionTime":"2025-10-14T15:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.407578 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.407623 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.407635 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.407654 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.407669 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:24Z","lastTransitionTime":"2025-10-14T15:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.510215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.510656 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.510836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.511060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.511205 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:24Z","lastTransitionTime":"2025-10-14T15:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.614685 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.614736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.614755 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.614778 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.614794 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:24Z","lastTransitionTime":"2025-10-14T15:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.717616 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.717674 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.717690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.717711 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.717726 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:24Z","lastTransitionTime":"2025-10-14T15:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.762904 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:24 crc kubenswrapper[4945]: E1014 15:15:24.763032 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.763155 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.762850 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:24 crc kubenswrapper[4945]: E1014 15:15:24.763389 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:24 crc kubenswrapper[4945]: E1014 15:15:24.763567 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.820562 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.820609 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.820622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.820641 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.820655 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:24Z","lastTransitionTime":"2025-10-14T15:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.923548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.923915 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.924016 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.924113 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.924208 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:24Z","lastTransitionTime":"2025-10-14T15:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.944633 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.955841 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.961687 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:24Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.976677 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:24Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:24 crc kubenswrapper[4945]: I1014 15:15:24.996368 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:24Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.010699 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.025040 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.026095 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.026165 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.026185 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.026212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.026229 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:25Z","lastTransitionTime":"2025-10-14T15:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.037779 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.054179 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.071319 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.089012 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.107006 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.123980 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.128376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.128424 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.128440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.128464 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.128479 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:25Z","lastTransitionTime":"2025-10-14T15:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.139997 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.156728 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.172020 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.186180 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.208523 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:02Z\\\",\\\"message\\\":\\\"calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z]\\\\nI1014 15:15:02.926934 6379 services_controller.go:434] Service openshift-machine-api/cluster-autoscaler-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{cluster-autoscaler-operator openshift-machine-api b8e0040a-0eca-4299-ac4a-f26a24879998 4394 0 2025-02-23 05:12:30 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:cluster-autoscaler-operator] map[exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:cluster-autoscaler-operator-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007ae224b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:25Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.231086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.231114 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.231123 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.231144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.231163 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:25Z","lastTransitionTime":"2025-10-14T15:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.334455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.334509 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.334529 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.334563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.334587 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:25Z","lastTransitionTime":"2025-10-14T15:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.437210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.437240 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.437252 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.437276 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.437299 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:25Z","lastTransitionTime":"2025-10-14T15:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.540154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.540229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.540256 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.540286 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.540311 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:25Z","lastTransitionTime":"2025-10-14T15:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.642621 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.642663 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.642675 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.642691 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.642704 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:25Z","lastTransitionTime":"2025-10-14T15:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.745322 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.745360 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.745368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.745386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.745395 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:25Z","lastTransitionTime":"2025-10-14T15:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.761937 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:25 crc kubenswrapper[4945]: E1014 15:15:25.762168 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.847823 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.847901 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.847915 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.847932 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.847946 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:25Z","lastTransitionTime":"2025-10-14T15:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.950448 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.950497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.950512 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.950532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:25 crc kubenswrapper[4945]: I1014 15:15:25.950547 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:25Z","lastTransitionTime":"2025-10-14T15:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.053869 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.053976 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.053994 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.054021 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.054041 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:26Z","lastTransitionTime":"2025-10-14T15:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.156599 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.156660 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.156677 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.156703 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.156721 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:26Z","lastTransitionTime":"2025-10-14T15:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.260072 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.260111 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.260119 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.260132 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.260141 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:26Z","lastTransitionTime":"2025-10-14T15:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.362274 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.362347 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.362370 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.362399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.362420 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:26Z","lastTransitionTime":"2025-10-14T15:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.465651 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.465691 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.465700 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.465715 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.465726 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:26Z","lastTransitionTime":"2025-10-14T15:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.568133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.568440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.568632 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.568813 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.569020 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:26Z","lastTransitionTime":"2025-10-14T15:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.671853 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.671945 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.671979 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.672010 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.672034 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:26Z","lastTransitionTime":"2025-10-14T15:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.762983 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:26 crc kubenswrapper[4945]: E1014 15:15:26.763131 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.762983 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:26 crc kubenswrapper[4945]: E1014 15:15:26.763212 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.763512 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:26 crc kubenswrapper[4945]: E1014 15:15:26.763794 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.774119 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.774165 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.774184 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.774205 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.774224 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:26Z","lastTransitionTime":"2025-10-14T15:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.876966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.877046 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.877073 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.877105 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.877128 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:26Z","lastTransitionTime":"2025-10-14T15:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.980551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.980622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.980638 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.980661 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:26 crc kubenswrapper[4945]: I1014 15:15:26.980675 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:26Z","lastTransitionTime":"2025-10-14T15:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.083517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.083585 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.083596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.083617 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.083630 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.186213 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.186265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.186282 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.186305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.186321 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.288646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.288689 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.288700 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.288719 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.288732 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.391452 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.391517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.391533 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.391558 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.391572 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.494009 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.494060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.494072 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.494090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.494102 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.597824 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.597866 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.597933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.597960 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.597977 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.644942 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.644986 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.645002 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.645022 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.645038 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: E1014 15:15:27.661166 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:27Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.666002 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.666056 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.666076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.666105 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.666127 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: E1014 15:15:27.682735 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:27Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.687313 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.687341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.687352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.687366 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.687377 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: E1014 15:15:27.699184 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:27Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.703377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.703425 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.703442 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.703463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.703478 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: E1014 15:15:27.722033 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:27Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.726267 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.726313 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.726329 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.726352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.726368 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: E1014 15:15:27.747613 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:27Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:27 crc kubenswrapper[4945]: E1014 15:15:27.747847 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.750200 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.750320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.750386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.750462 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.750561 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.762616 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:27 crc kubenswrapper[4945]: E1014 15:15:27.763298 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.763666 4945 scope.go:117] "RemoveContainer" containerID="3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.853706 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.854286 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.854380 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.854539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.854677 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.957700 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.957741 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.957753 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.957769 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:27 crc kubenswrapper[4945]: I1014 15:15:27.957781 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:27Z","lastTransitionTime":"2025-10-14T15:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.060133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.060181 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.060196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.060212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.060224 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:28Z","lastTransitionTime":"2025-10-14T15:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.104685 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/1.log" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.106798 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerStarted","Data":"99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03"} Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.107968 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.118712 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.128243 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.143585 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.158579 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.161968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.162003 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.162012 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.162025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.162034 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:28Z","lastTransitionTime":"2025-10-14T15:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.171417 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.188619 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.202965 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.223943 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.242286 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.255302 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.264629 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.264670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.264678 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.264694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.264704 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:28Z","lastTransitionTime":"2025-10-14T15:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.269670 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.280484 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.290301 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5fbb433-31b8-4b3c-ae3d-6d34d62cde7f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.301266 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.318732 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:02Z\\\",\\\"message\\\":\\\"calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z]\\\\nI1014 15:15:02.926934 6379 services_controller.go:434] Service openshift-machine-api/cluster-autoscaler-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{cluster-autoscaler-operator openshift-machine-api b8e0040a-0eca-4299-ac4a-f26a24879998 4394 0 2025-02-23 05:12:30 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:cluster-autoscaler-operator] map[exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:cluster-autoscaler-operator-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007ae224b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.331105 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.340338 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:28Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.367224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.367260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.367269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.367282 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.367292 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:28Z","lastTransitionTime":"2025-10-14T15:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.470150 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.470190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.470199 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.470213 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.470222 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:28Z","lastTransitionTime":"2025-10-14T15:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.573844 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.574203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.574369 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.574510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.574637 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:28Z","lastTransitionTime":"2025-10-14T15:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.677686 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.677741 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.677754 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.677772 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.677788 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:28Z","lastTransitionTime":"2025-10-14T15:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.763085 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.763143 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.763199 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:28 crc kubenswrapper[4945]: E1014 15:15:28.763311 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:28 crc kubenswrapper[4945]: E1014 15:15:28.763483 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:28 crc kubenswrapper[4945]: E1014 15:15:28.763515 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.780496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.780551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.780571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.780599 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.780617 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:28Z","lastTransitionTime":"2025-10-14T15:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.882788 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.882851 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.882863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.882939 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.882966 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:28Z","lastTransitionTime":"2025-10-14T15:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.985791 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.986193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.986369 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.986592 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:28 crc kubenswrapper[4945]: I1014 15:15:28.986809 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:28Z","lastTransitionTime":"2025-10-14T15:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.092953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.092998 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.093010 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.093032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.093044 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:29Z","lastTransitionTime":"2025-10-14T15:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.110169 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/2.log" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.110843 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/1.log" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.113503 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03" exitCode=1 Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.113532 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03"} Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.113561 4945 scope.go:117] "RemoveContainer" containerID="3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.114258 4945 scope.go:117] "RemoveContainer" containerID="99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03" Oct 14 15:15:29 crc kubenswrapper[4945]: E1014 15:15:29.114465 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.128371 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5fbb433-31b8-4b3c-ae3d-6d34d62cde7f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.141082 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.153914 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.171385 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.182179 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.195380 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.195408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.195421 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.195438 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.195451 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:29Z","lastTransitionTime":"2025-10-14T15:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.200607 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3828cc9a2975aa5bc6b61f9f359797a7d811df96c00fe7427386fecfc04921e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:02Z\\\",\\\"message\\\":\\\"calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:02Z is after 2025-08-24T17:21:41Z]\\\\nI1014 15:15:02.926934 6379 services_controller.go:434] Service openshift-machine-api/cluster-autoscaler-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{cluster-autoscaler-operator openshift-machine-api b8e0040a-0eca-4299-ac4a-f26a24879998 4394 0 2025-02-23 05:12:30 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:cluster-autoscaler-operator] map[exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:cluster-autoscaler-operator-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007ae224b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:metri\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:28Z\\\",\\\"message\\\":\\\"openshift-kube-storage-version-migrator-operator for network=default : 1.541893ms\\\\nI1014 15:15:28.613080 6660 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613130 6660 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613145 6660 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1014 15:15:28.613163 6660 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.728948ms\\\\nI1014 15:15:28.612485 6660 obj_re\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.211357 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.224628 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.244706 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.257574 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.271545 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.281554 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.290683 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.298510 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.298554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.298569 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.298590 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.298605 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:29Z","lastTransitionTime":"2025-10-14T15:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.301309 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.312761 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.327875 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.339470 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:29Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.401725 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.401766 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.401778 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.401794 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.401806 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:29Z","lastTransitionTime":"2025-10-14T15:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.503935 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.504014 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.504034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.504058 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.504076 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:29Z","lastTransitionTime":"2025-10-14T15:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.607368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.607426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.607443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.607467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.607484 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:29Z","lastTransitionTime":"2025-10-14T15:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.710780 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.710839 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.710856 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.710912 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.710936 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:29Z","lastTransitionTime":"2025-10-14T15:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.761905 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:29 crc kubenswrapper[4945]: E1014 15:15:29.762143 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.813357 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.813431 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.813453 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.813480 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.813500 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:29Z","lastTransitionTime":"2025-10-14T15:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.917055 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.917112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.917132 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.917155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:29 crc kubenswrapper[4945]: I1014 15:15:29.917172 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:29Z","lastTransitionTime":"2025-10-14T15:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.019789 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.019848 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.019858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.020103 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.020129 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:30Z","lastTransitionTime":"2025-10-14T15:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.122754 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/2.log" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.125814 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.126089 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.126121 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.126152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.126175 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:30Z","lastTransitionTime":"2025-10-14T15:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.128858 4945 scope.go:117] "RemoveContainer" containerID="99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03" Oct 14 15:15:30 crc kubenswrapper[4945]: E1014 15:15:30.129473 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.152002 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.171708 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5fbb433-31b8-4b3c-ae3d-6d34d62cde7f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.192258 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.217396 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:28Z\\\",\\\"message\\\":\\\"openshift-kube-storage-version-migrator-operator for network=default : 1.541893ms\\\\nI1014 15:15:28.613080 6660 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613130 6660 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613145 6660 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1014 15:15:28.613163 6660 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.728948ms\\\\nI1014 15:15:28.612485 6660 obj_re\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.229145 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.229198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.229213 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.229230 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.229242 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:30Z","lastTransitionTime":"2025-10-14T15:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.235442 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.247468 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.259309 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.270733 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.282971 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.296000 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.305559 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.316240 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.328763 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.331537 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.331582 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.331596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.331611 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.331644 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:30Z","lastTransitionTime":"2025-10-14T15:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.339330 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.352969 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.364227 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.374758 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.434671 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.434729 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.434747 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.434773 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.434790 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:30Z","lastTransitionTime":"2025-10-14T15:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.537557 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.537608 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.537625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.537649 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.537665 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:30Z","lastTransitionTime":"2025-10-14T15:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.641098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.641152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.641166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.641186 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.641197 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:30Z","lastTransitionTime":"2025-10-14T15:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.744297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.744377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.744400 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.744432 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.744457 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:30Z","lastTransitionTime":"2025-10-14T15:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.761911 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.761992 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:30 crc kubenswrapper[4945]: E1014 15:15:30.762071 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:30 crc kubenswrapper[4945]: E1014 15:15:30.762188 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.762217 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:30 crc kubenswrapper[4945]: E1014 15:15:30.762378 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.786427 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.803553 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.817849 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.834797 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.847504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.847559 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.847579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.847604 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.847621 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:30Z","lastTransitionTime":"2025-10-14T15:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.857629 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.873066 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.885095 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5fbb433-31b8-4b3c-ae3d-6d34d62cde7f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.901531 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.919396 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.937156 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.950694 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.950736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.950746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.950761 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.950770 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:30Z","lastTransitionTime":"2025-10-14T15:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.952704 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.976672 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:28Z\\\",\\\"message\\\":\\\"openshift-kube-storage-version-migrator-operator for network=default : 1.541893ms\\\\nI1014 15:15:28.613080 6660 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613130 6660 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613145 6660 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1014 15:15:28.613163 6660 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.728948ms\\\\nI1014 15:15:28.612485 6660 obj_re\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:30 crc kubenswrapper[4945]: I1014 15:15:30.995658 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:30Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.012138 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:31Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.024279 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:31Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.040436 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:31Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.052801 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.052853 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.052866 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.052912 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.052966 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:31Z","lastTransitionTime":"2025-10-14T15:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.053955 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:31Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.155677 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.155737 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.155755 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.155776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.155791 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:31Z","lastTransitionTime":"2025-10-14T15:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.258459 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.258516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.258528 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.258547 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.258560 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:31Z","lastTransitionTime":"2025-10-14T15:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.361795 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.361871 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.361939 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.361970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.361994 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:31Z","lastTransitionTime":"2025-10-14T15:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.464540 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.464660 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.464685 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.464715 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.464734 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:31Z","lastTransitionTime":"2025-10-14T15:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.567682 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.567737 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.567755 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.567781 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.567798 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:31Z","lastTransitionTime":"2025-10-14T15:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.671082 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.671149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.671168 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.671199 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.671224 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:31Z","lastTransitionTime":"2025-10-14T15:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.762937 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:31 crc kubenswrapper[4945]: E1014 15:15:31.763075 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.773261 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.773289 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.773296 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.773308 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.773318 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:31Z","lastTransitionTime":"2025-10-14T15:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.876309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.876377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.876395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.876422 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.876449 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:31Z","lastTransitionTime":"2025-10-14T15:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.979244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.979723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.980037 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.981580 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:31 crc kubenswrapper[4945]: I1014 15:15:31.981755 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:31Z","lastTransitionTime":"2025-10-14T15:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.084695 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.084810 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.084829 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.084856 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.084899 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:32Z","lastTransitionTime":"2025-10-14T15:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.188195 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.188304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.188326 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.188350 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.188368 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:32Z","lastTransitionTime":"2025-10-14T15:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.292235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.292286 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.292306 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.292330 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.292347 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:32Z","lastTransitionTime":"2025-10-14T15:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.394904 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.394944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.394958 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.394975 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.394988 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:32Z","lastTransitionTime":"2025-10-14T15:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.497556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.497622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.497640 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.497667 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.497684 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:32Z","lastTransitionTime":"2025-10-14T15:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.601366 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.601714 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.601738 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.601766 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.601786 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:32Z","lastTransitionTime":"2025-10-14T15:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.704932 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.704984 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.705001 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.705027 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.705047 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:32Z","lastTransitionTime":"2025-10-14T15:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.762724 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.762795 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:32 crc kubenswrapper[4945]: E1014 15:15:32.762932 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.763004 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:32 crc kubenswrapper[4945]: E1014 15:15:32.763146 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:32 crc kubenswrapper[4945]: E1014 15:15:32.763415 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.809223 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.809341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.809358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.809375 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.809387 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:32Z","lastTransitionTime":"2025-10-14T15:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.912193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.912240 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.912257 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.912283 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:32 crc kubenswrapper[4945]: I1014 15:15:32.912303 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:32Z","lastTransitionTime":"2025-10-14T15:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.014568 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.014609 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.014619 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.014635 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.014646 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:33Z","lastTransitionTime":"2025-10-14T15:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.116911 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.116959 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.116977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.117000 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.117018 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:33Z","lastTransitionTime":"2025-10-14T15:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.220023 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.220088 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.220111 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.220150 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.220185 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:33Z","lastTransitionTime":"2025-10-14T15:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.322760 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.322841 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.322868 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.322945 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.322973 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:33Z","lastTransitionTime":"2025-10-14T15:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.425574 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.425607 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.425615 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.425628 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.425637 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:33Z","lastTransitionTime":"2025-10-14T15:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.527616 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.527654 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.527662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.527676 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.527685 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:33Z","lastTransitionTime":"2025-10-14T15:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.630377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.630417 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.630426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.630443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.630452 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:33Z","lastTransitionTime":"2025-10-14T15:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.733301 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.733360 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.733377 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.733401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.733417 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:33Z","lastTransitionTime":"2025-10-14T15:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.761950 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:33 crc kubenswrapper[4945]: E1014 15:15:33.762177 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.836129 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.836187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.836205 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.836230 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.836251 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:33Z","lastTransitionTime":"2025-10-14T15:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.939448 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.939485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.939494 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.939506 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:33 crc kubenswrapper[4945]: I1014 15:15:33.939517 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:33Z","lastTransitionTime":"2025-10-14T15:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.042186 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.042242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.042256 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.042274 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.042285 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:34Z","lastTransitionTime":"2025-10-14T15:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.144137 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.144183 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.144195 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.144213 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.144224 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:34Z","lastTransitionTime":"2025-10-14T15:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.246996 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.247062 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.247082 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.247104 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.247122 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:34Z","lastTransitionTime":"2025-10-14T15:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.349318 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.349376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.349396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.349420 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.349437 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:34Z","lastTransitionTime":"2025-10-14T15:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.452231 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.452287 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.452297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.452311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.452322 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:34Z","lastTransitionTime":"2025-10-14T15:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.555310 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.555372 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.555384 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.555405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.555437 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:34Z","lastTransitionTime":"2025-10-14T15:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.658525 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.658584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.658609 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.658626 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.658638 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:34Z","lastTransitionTime":"2025-10-14T15:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.761969 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.762026 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.761975 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:34 crc kubenswrapper[4945]: E1014 15:15:34.762153 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:34 crc kubenswrapper[4945]: E1014 15:15:34.762285 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:34 crc kubenswrapper[4945]: E1014 15:15:34.762397 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.762503 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.762567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.762584 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.762605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.762621 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:34Z","lastTransitionTime":"2025-10-14T15:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.865504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.865572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.865594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.865622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.865644 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:34Z","lastTransitionTime":"2025-10-14T15:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.968494 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.968728 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.968771 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.968804 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:34 crc kubenswrapper[4945]: I1014 15:15:34.968829 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:34Z","lastTransitionTime":"2025-10-14T15:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.071717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.071782 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.071800 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.071825 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.071843 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:35Z","lastTransitionTime":"2025-10-14T15:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.174773 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.174819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.174836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.174858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.174902 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:35Z","lastTransitionTime":"2025-10-14T15:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.278016 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.278082 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.278095 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.278112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.278123 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:35Z","lastTransitionTime":"2025-10-14T15:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.380456 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.380512 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.380530 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.380553 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.380569 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:35Z","lastTransitionTime":"2025-10-14T15:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.484673 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.484825 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.484839 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.484856 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.484965 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:35Z","lastTransitionTime":"2025-10-14T15:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.588052 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.588137 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.588160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.588194 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.588218 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:35Z","lastTransitionTime":"2025-10-14T15:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.690822 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.690871 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.690901 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.690918 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.690931 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:35Z","lastTransitionTime":"2025-10-14T15:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.762641 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:35 crc kubenswrapper[4945]: E1014 15:15:35.762759 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.793396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.793432 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.793441 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.793457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.793469 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:35Z","lastTransitionTime":"2025-10-14T15:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.896483 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.896532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.896545 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.896563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.896580 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:35Z","lastTransitionTime":"2025-10-14T15:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.998827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.998866 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.998888 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.998902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:35 crc kubenswrapper[4945]: I1014 15:15:35.998913 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:35Z","lastTransitionTime":"2025-10-14T15:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.101298 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.101356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.101375 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.101393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.101403 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:36Z","lastTransitionTime":"2025-10-14T15:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.203828 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.203921 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.203933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.203951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.203962 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:36Z","lastTransitionTime":"2025-10-14T15:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.306507 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.306553 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.306565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.306590 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.306601 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:36Z","lastTransitionTime":"2025-10-14T15:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.408725 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.408764 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.408775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.408792 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.408804 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:36Z","lastTransitionTime":"2025-10-14T15:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.510975 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.511017 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.511027 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.511041 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.511051 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:36Z","lastTransitionTime":"2025-10-14T15:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.613577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.613670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.613696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.613723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.613740 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:36Z","lastTransitionTime":"2025-10-14T15:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.717053 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.717113 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.717127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.717152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.717167 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:36Z","lastTransitionTime":"2025-10-14T15:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.762496 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.762535 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:36 crc kubenswrapper[4945]: E1014 15:15:36.762689 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.762906 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:36 crc kubenswrapper[4945]: E1014 15:15:36.762979 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:36 crc kubenswrapper[4945]: E1014 15:15:36.763145 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.819067 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.819114 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.819129 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.819150 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.819163 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:36Z","lastTransitionTime":"2025-10-14T15:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.926336 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.926381 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.926393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.926410 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:36 crc kubenswrapper[4945]: I1014 15:15:36.926425 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:36Z","lastTransitionTime":"2025-10-14T15:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.029309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.029349 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.029358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.029374 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.029384 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:37Z","lastTransitionTime":"2025-10-14T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.131994 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.132034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.132045 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.132060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.132072 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:37Z","lastTransitionTime":"2025-10-14T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.234570 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.234615 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.234624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.234639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.234651 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:37Z","lastTransitionTime":"2025-10-14T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.337337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.337392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.337405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.337423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.337436 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:37Z","lastTransitionTime":"2025-10-14T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.440032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.440066 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.440087 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.440101 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.440111 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:37Z","lastTransitionTime":"2025-10-14T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.542076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.542105 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.542127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.542144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.542155 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:37Z","lastTransitionTime":"2025-10-14T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.644348 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.644376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.644387 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.644401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.644413 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:37Z","lastTransitionTime":"2025-10-14T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.746308 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.746380 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.746405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.746435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.746457 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:37Z","lastTransitionTime":"2025-10-14T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.762906 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:37 crc kubenswrapper[4945]: E1014 15:15:37.763180 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.848626 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.848973 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.849082 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.849177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.849268 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:37Z","lastTransitionTime":"2025-10-14T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.904165 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:37 crc kubenswrapper[4945]: E1014 15:15:37.904328 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:37 crc kubenswrapper[4945]: E1014 15:15:37.904572 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs podName:be17bc84-909e-4150-8ee7-9a7d23173ba0 nodeName:}" failed. No retries permitted until 2025-10-14 15:16:09.904553142 +0000 UTC m=+99.888601510 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs") pod "network-metrics-daemon-ftg5h" (UID: "be17bc84-909e-4150-8ee7-9a7d23173ba0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.951411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.951456 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.951468 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.951483 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.951494 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:37Z","lastTransitionTime":"2025-10-14T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.995816 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.995889 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.995902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.995926 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:37 crc kubenswrapper[4945]: I1014 15:15:37.995939 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:37Z","lastTransitionTime":"2025-10-14T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: E1014 15:15:38.014703 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:38Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.018330 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.018432 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.018495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.018556 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.018611 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: E1014 15:15:38.031895 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:38Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.056200 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.056258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.056278 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.056305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.056323 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: E1014 15:15:38.078656 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:38Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.083401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.083605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.083714 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.083826 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.083939 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: E1014 15:15:38.098907 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:38Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.103030 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.103293 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.103419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.103548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.103660 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: E1014 15:15:38.116756 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:38Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:38 crc kubenswrapper[4945]: E1014 15:15:38.117275 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.119241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.119276 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.119284 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.119298 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.119310 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.222198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.222233 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.222242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.222256 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.222265 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.324704 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.324777 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.324792 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.324812 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.324824 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.427129 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.427211 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.427232 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.427260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.427288 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.529043 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.529084 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.529097 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.529112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.529123 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.645983 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.646029 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.646043 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.646060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.646072 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.748758 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.748804 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.748814 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.748832 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.748844 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.762125 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.762143 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.762144 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:38 crc kubenswrapper[4945]: E1014 15:15:38.762513 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:38 crc kubenswrapper[4945]: E1014 15:15:38.762602 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:38 crc kubenswrapper[4945]: E1014 15:15:38.762290 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.851059 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.851102 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.851114 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.851131 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.851144 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.952820 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.952848 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.952858 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.952896 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:38 crc kubenswrapper[4945]: I1014 15:15:38.952913 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:38Z","lastTransitionTime":"2025-10-14T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.055973 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.056011 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.056020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.056034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.056043 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:39Z","lastTransitionTime":"2025-10-14T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.157509 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.157594 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.157612 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.157636 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.157653 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:39Z","lastTransitionTime":"2025-10-14T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.259938 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.259971 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.259982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.259997 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.260008 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:39Z","lastTransitionTime":"2025-10-14T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.361806 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.362283 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.362503 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.362700 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.362911 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:39Z","lastTransitionTime":"2025-10-14T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.465720 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.465798 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.465820 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.465849 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.465905 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:39Z","lastTransitionTime":"2025-10-14T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.568662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.568751 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.568779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.568813 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.568830 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:39Z","lastTransitionTime":"2025-10-14T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.672096 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.672146 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.672157 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.672178 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.672203 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:39Z","lastTransitionTime":"2025-10-14T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.762102 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:39 crc kubenswrapper[4945]: E1014 15:15:39.762262 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.774050 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.774090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.774099 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.774112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.774120 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:39Z","lastTransitionTime":"2025-10-14T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.876203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.876229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.876238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.876250 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.876260 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:39Z","lastTransitionTime":"2025-10-14T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.978084 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.978119 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.978130 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.978146 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:39 crc kubenswrapper[4945]: I1014 15:15:39.978157 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:39Z","lastTransitionTime":"2025-10-14T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.080577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.080611 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.080622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.080637 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.080649 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:40Z","lastTransitionTime":"2025-10-14T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.158923 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b76xs_f82b9cd9-7a14-4dd9-bb0e-fde942389666/kube-multus/0.log" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.158979 4945 generic.go:334] "Generic (PLEG): container finished" podID="f82b9cd9-7a14-4dd9-bb0e-fde942389666" containerID="42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289" exitCode=1 Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.159016 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b76xs" event={"ID":"f82b9cd9-7a14-4dd9-bb0e-fde942389666","Type":"ContainerDied","Data":"42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289"} Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.159399 4945 scope.go:117] "RemoveContainer" containerID="42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.175458 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.183650 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.183674 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.183689 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.183710 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.183721 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:40Z","lastTransitionTime":"2025-10-14T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.189539 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.210727 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:28Z\\\",\\\"message\\\":\\\"openshift-kube-storage-version-migrator-operator for network=default : 1.541893ms\\\\nI1014 15:15:28.613080 6660 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613130 6660 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613145 6660 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1014 15:15:28.613163 6660 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.728948ms\\\\nI1014 15:15:28.612485 6660 obj_re\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.224616 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.237040 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.249537 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.259798 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.274017 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.286337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.286373 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.286391 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.286410 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.286423 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:40Z","lastTransitionTime":"2025-10-14T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.290068 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.300804 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.311510 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.321190 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.329513 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.342583 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.354036 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5fbb433-31b8-4b3c-ae3d-6d34d62cde7f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.365232 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.378503 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:39Z\\\",\\\"message\\\":\\\"2025-10-14T15:14:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0\\\\n2025-10-14T15:14:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0 to /host/opt/cni/bin/\\\\n2025-10-14T15:14:54Z [verbose] multus-daemon started\\\\n2025-10-14T15:14:54Z [verbose] Readiness Indicator file check\\\\n2025-10-14T15:15:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.391900 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.391951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.391966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.391985 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.391998 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:40Z","lastTransitionTime":"2025-10-14T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.494597 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.494639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.494651 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.494705 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.494716 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:40Z","lastTransitionTime":"2025-10-14T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.597435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.597475 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.597484 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.597498 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.597511 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:40Z","lastTransitionTime":"2025-10-14T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.699669 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.699697 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.699705 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.699718 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.699726 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:40Z","lastTransitionTime":"2025-10-14T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.762012 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.762104 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:40 crc kubenswrapper[4945]: E1014 15:15:40.762151 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.762012 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:40 crc kubenswrapper[4945]: E1014 15:15:40.762228 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:40 crc kubenswrapper[4945]: E1014 15:15:40.762300 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.779908 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.792368 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.801380 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.801403 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.801412 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.801426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.801437 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:40Z","lastTransitionTime":"2025-10-14T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.814650 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:28Z\\\",\\\"message\\\":\\\"openshift-kube-storage-version-migrator-operator for network=default : 1.541893ms\\\\nI1014 15:15:28.613080 6660 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613130 6660 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613145 6660 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1014 15:15:28.613163 6660 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.728948ms\\\\nI1014 15:15:28.612485 6660 obj_re\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.832294 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.848326 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.860893 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.873369 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.884365 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.896442 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.903953 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.903994 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.904004 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.904020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.904031 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:40Z","lastTransitionTime":"2025-10-14T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.907032 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.917626 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.927452 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.936406 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.946859 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.957987 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5fbb433-31b8-4b3c-ae3d-6d34d62cde7f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.967989 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:40 crc kubenswrapper[4945]: I1014 15:15:40.980369 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:39Z\\\",\\\"message\\\":\\\"2025-10-14T15:14:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0\\\\n2025-10-14T15:14:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0 to /host/opt/cni/bin/\\\\n2025-10-14T15:14:54Z [verbose] multus-daemon started\\\\n2025-10-14T15:14:54Z [verbose] Readiness Indicator file check\\\\n2025-10-14T15:15:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:40Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.005733 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.005781 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.005794 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.005814 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.005828 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:41Z","lastTransitionTime":"2025-10-14T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.107320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.107349 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.107361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.107378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.107389 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:41Z","lastTransitionTime":"2025-10-14T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.163265 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b76xs_f82b9cd9-7a14-4dd9-bb0e-fde942389666/kube-multus/0.log" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.163323 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b76xs" event={"ID":"f82b9cd9-7a14-4dd9-bb0e-fde942389666","Type":"ContainerStarted","Data":"17eb13e41970758fd540b4fff10d1848e4f0da616768856063641a4d1994340c"} Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.179058 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.193698 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.209301 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.209996 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.210017 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.210025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.210038 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.210046 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:41Z","lastTransitionTime":"2025-10-14T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.218926 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.229516 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.239515 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.250761 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.262315 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.275806 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.287708 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.297943 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.309090 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5fbb433-31b8-4b3c-ae3d-6d34d62cde7f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.312310 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.312357 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.312370 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.312389 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.312404 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:41Z","lastTransitionTime":"2025-10-14T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.319853 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.338572 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb13e41970758fd540b4fff10d1848e4f0da616768856063641a4d1994340c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:39Z\\\",\\\"message\\\":\\\"2025-10-14T15:14:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0\\\\n2025-10-14T15:14:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0 to /host/opt/cni/bin/\\\\n2025-10-14T15:14:54Z [verbose] multus-daemon started\\\\n2025-10-14T15:14:54Z [verbose] Readiness Indicator file check\\\\n2025-10-14T15:15:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.352203 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.361224 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.377094 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:28Z\\\",\\\"message\\\":\\\"openshift-kube-storage-version-migrator-operator for network=default : 1.541893ms\\\\nI1014 15:15:28.613080 6660 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613130 6660 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613145 6660 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1014 15:15:28.613163 6660 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.728948ms\\\\nI1014 15:15:28.612485 6660 obj_re\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:41Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.414268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.414303 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.414315 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.414331 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.414343 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:41Z","lastTransitionTime":"2025-10-14T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.516461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.516514 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.516522 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.516536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.516547 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:41Z","lastTransitionTime":"2025-10-14T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.619244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.619297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.619310 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.619329 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.619343 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:41Z","lastTransitionTime":"2025-10-14T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.721867 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.721929 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.721941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.721959 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.721974 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:41Z","lastTransitionTime":"2025-10-14T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.762749 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:41 crc kubenswrapper[4945]: E1014 15:15:41.762906 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.824586 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.824643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.824655 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.824671 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.824682 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:41Z","lastTransitionTime":"2025-10-14T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.927060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.927099 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.927113 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.927131 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:41 crc kubenswrapper[4945]: I1014 15:15:41.927145 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:41Z","lastTransitionTime":"2025-10-14T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.029617 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.029672 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.029690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.029714 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.029732 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:42Z","lastTransitionTime":"2025-10-14T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.132318 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.132374 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.132391 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.132413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.132430 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:42Z","lastTransitionTime":"2025-10-14T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.234135 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.234159 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.234166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.234178 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.234186 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:42Z","lastTransitionTime":"2025-10-14T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.335952 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.336024 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.336041 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.336088 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.336104 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:42Z","lastTransitionTime":"2025-10-14T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.438902 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.438950 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.438977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.438993 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.439004 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:42Z","lastTransitionTime":"2025-10-14T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.540719 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.540773 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.540813 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.540831 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.540845 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:42Z","lastTransitionTime":"2025-10-14T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.642861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.642907 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.642915 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.642928 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.642937 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:42Z","lastTransitionTime":"2025-10-14T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.744479 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.744541 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.744550 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.744563 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.744572 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:42Z","lastTransitionTime":"2025-10-14T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.762340 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:42 crc kubenswrapper[4945]: E1014 15:15:42.762460 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.762623 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:42 crc kubenswrapper[4945]: E1014 15:15:42.762673 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.762772 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:42 crc kubenswrapper[4945]: E1014 15:15:42.762933 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.763524 4945 scope.go:117] "RemoveContainer" containerID="99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03" Oct 14 15:15:42 crc kubenswrapper[4945]: E1014 15:15:42.763692 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.846716 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.846754 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.846763 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.846778 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.846787 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:42Z","lastTransitionTime":"2025-10-14T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.948500 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.948543 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.948554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.948571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:42 crc kubenswrapper[4945]: I1014 15:15:42.948582 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:42Z","lastTransitionTime":"2025-10-14T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.051094 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.051132 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.051142 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.051157 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.051168 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:43Z","lastTransitionTime":"2025-10-14T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.156597 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.156646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.156662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.156678 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.156688 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:43Z","lastTransitionTime":"2025-10-14T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.259008 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.259051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.259061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.259078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.259090 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:43Z","lastTransitionTime":"2025-10-14T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.360595 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.360634 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.360645 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.360661 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.360671 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:43Z","lastTransitionTime":"2025-10-14T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.462746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.462779 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.462789 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.462802 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.462810 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:43Z","lastTransitionTime":"2025-10-14T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.565824 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.565868 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.565899 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.565917 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.565929 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:43Z","lastTransitionTime":"2025-10-14T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.668304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.668822 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.668926 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.669031 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.669115 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:43Z","lastTransitionTime":"2025-10-14T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.762180 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:43 crc kubenswrapper[4945]: E1014 15:15:43.762307 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.772923 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.772959 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.772968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.772982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.772994 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:43Z","lastTransitionTime":"2025-10-14T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.874684 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.874735 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.874748 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.874767 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.874780 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:43Z","lastTransitionTime":"2025-10-14T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.977213 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.977275 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.977289 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.977309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:43 crc kubenswrapper[4945]: I1014 15:15:43.977324 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:43Z","lastTransitionTime":"2025-10-14T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.081290 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.081325 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.081338 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.081384 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.081397 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:44Z","lastTransitionTime":"2025-10-14T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.184297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.184365 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.184393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.184418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.184441 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:44Z","lastTransitionTime":"2025-10-14T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.286726 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.286808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.286825 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.286847 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.286860 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:44Z","lastTransitionTime":"2025-10-14T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.389242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.389293 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.389304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.389322 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.389335 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:44Z","lastTransitionTime":"2025-10-14T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.491968 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.492049 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.492061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.492076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.492086 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:44Z","lastTransitionTime":"2025-10-14T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.595020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.595225 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.595256 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.595279 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.595295 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:44Z","lastTransitionTime":"2025-10-14T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.697709 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.697749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.697761 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.697776 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.697789 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:44Z","lastTransitionTime":"2025-10-14T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.761992 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.762195 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:44 crc kubenswrapper[4945]: E1014 15:15:44.762205 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:44 crc kubenswrapper[4945]: E1014 15:15:44.762343 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.762536 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:44 crc kubenswrapper[4945]: E1014 15:15:44.762626 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.799913 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.800113 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.800168 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.800259 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.800317 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:44Z","lastTransitionTime":"2025-10-14T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.903777 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.903862 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.903929 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.903961 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:44 crc kubenswrapper[4945]: I1014 15:15:44.904021 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:44Z","lastTransitionTime":"2025-10-14T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.006466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.006506 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.006517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.006534 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.006546 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:45Z","lastTransitionTime":"2025-10-14T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.108759 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.108810 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.108827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.108848 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.108865 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:45Z","lastTransitionTime":"2025-10-14T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.211332 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.211367 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.211380 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.211396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.211408 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:45Z","lastTransitionTime":"2025-10-14T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.313892 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.313982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.314001 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.314023 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.314032 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:45Z","lastTransitionTime":"2025-10-14T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.417964 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.418054 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.418684 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.418727 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.418745 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:45Z","lastTransitionTime":"2025-10-14T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.548096 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.548147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.548157 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.548173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.548183 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:45Z","lastTransitionTime":"2025-10-14T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.650149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.650204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.650221 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.650263 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.650280 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:45Z","lastTransitionTime":"2025-10-14T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.753203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.753250 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.753267 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.753289 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.753306 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:45Z","lastTransitionTime":"2025-10-14T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.763005 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:45 crc kubenswrapper[4945]: E1014 15:15:45.763135 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.856210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.856264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.856282 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.856304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.856322 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:45Z","lastTransitionTime":"2025-10-14T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.958971 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.959011 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.959019 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.959064 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:45 crc kubenswrapper[4945]: I1014 15:15:45.959074 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:45Z","lastTransitionTime":"2025-10-14T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.063943 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.064012 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.064034 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.064063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.064085 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:46Z","lastTransitionTime":"2025-10-14T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.168051 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.168096 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.168108 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.168128 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.168139 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:46Z","lastTransitionTime":"2025-10-14T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.271358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.271384 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.271392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.271404 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.271413 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:46Z","lastTransitionTime":"2025-10-14T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.374198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.374244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.374258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.374277 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.374290 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:46Z","lastTransitionTime":"2025-10-14T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.476715 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.476766 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.476781 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.476800 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.476813 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:46Z","lastTransitionTime":"2025-10-14T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.578746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.578794 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.578808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.578830 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.578844 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:46Z","lastTransitionTime":"2025-10-14T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.681680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.681940 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.682092 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.682122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.682132 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:46Z","lastTransitionTime":"2025-10-14T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.762275 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.762328 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:46 crc kubenswrapper[4945]: E1014 15:15:46.762403 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.762497 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:46 crc kubenswrapper[4945]: E1014 15:15:46.762620 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:46 crc kubenswrapper[4945]: E1014 15:15:46.762682 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.784124 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.784156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.784168 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.784184 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.784194 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:46Z","lastTransitionTime":"2025-10-14T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.886395 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.886426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.886436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.886451 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.886462 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:46Z","lastTransitionTime":"2025-10-14T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.989997 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.990121 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.990205 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.990266 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:46 crc kubenswrapper[4945]: I1014 15:15:46.990287 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:46Z","lastTransitionTime":"2025-10-14T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.092699 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.092734 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.092746 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.092763 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.092775 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:47Z","lastTransitionTime":"2025-10-14T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.203916 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.203945 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.203958 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.203982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.203995 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:47Z","lastTransitionTime":"2025-10-14T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.305952 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.306015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.306037 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.306065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.306088 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:47Z","lastTransitionTime":"2025-10-14T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.408082 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.408122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.408136 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.408152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.408165 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:47Z","lastTransitionTime":"2025-10-14T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.509686 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.509734 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.509750 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.509770 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.509782 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:47Z","lastTransitionTime":"2025-10-14T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.611967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.612016 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.612027 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.612049 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.612059 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:47Z","lastTransitionTime":"2025-10-14T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.714109 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.714150 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.714162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.714196 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.714209 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:47Z","lastTransitionTime":"2025-10-14T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.761959 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:47 crc kubenswrapper[4945]: E1014 15:15:47.762170 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.816596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.816652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.816676 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.816705 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.816726 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:47Z","lastTransitionTime":"2025-10-14T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.919851 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.919958 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.919992 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.920022 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:47 crc kubenswrapper[4945]: I1014 15:15:47.920044 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:47Z","lastTransitionTime":"2025-10-14T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.022805 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.022867 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.022916 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.022941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.022958 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.126479 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.126536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.126554 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.126577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.126594 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.228972 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.229027 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.229047 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.229070 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.229086 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.263409 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.263478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.263497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.263522 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.263539 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: E1014 15:15:48.280248 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:48Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.284836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.284999 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.285022 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.285047 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.285064 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: E1014 15:15:48.308185 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:48Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.314722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.314768 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.314781 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.314801 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.314816 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: E1014 15:15:48.332767 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:48Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.336318 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.336379 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.336393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.336408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.336417 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: E1014 15:15:48.353030 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:48Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.358133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.358166 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.358175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.358188 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.358197 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: E1014 15:15:48.375310 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:48Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:48 crc kubenswrapper[4945]: E1014 15:15:48.376001 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.378268 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.378565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.378786 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.379061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.379265 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.481808 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.481906 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.481933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.481964 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.481985 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.585711 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.586091 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.586260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.586405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.586539 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.689149 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.689212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.689230 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.689254 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.689273 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.762176 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.762413 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.762321 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:48 crc kubenswrapper[4945]: E1014 15:15:48.762576 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:48 crc kubenswrapper[4945]: E1014 15:15:48.762564 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:48 crc kubenswrapper[4945]: E1014 15:15:48.762743 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.791536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.791616 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.791643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.791672 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.791696 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.894376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.894408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.894416 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.894430 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.894441 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.997127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.997190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.997208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.997235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:48 crc kubenswrapper[4945]: I1014 15:15:48.997254 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:48Z","lastTransitionTime":"2025-10-14T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.100406 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.100474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.100503 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.100534 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.100552 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:49Z","lastTransitionTime":"2025-10-14T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.204133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.204208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.204234 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.204262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.204279 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:49Z","lastTransitionTime":"2025-10-14T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.307038 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.307112 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.307131 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.307156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.307174 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:49Z","lastTransitionTime":"2025-10-14T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.409341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.409409 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.409428 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.409453 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.409471 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:49Z","lastTransitionTime":"2025-10-14T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.513151 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.513599 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.513828 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.514102 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.514378 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:49Z","lastTransitionTime":"2025-10-14T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.619103 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.619154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.619170 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.619194 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.619211 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:49Z","lastTransitionTime":"2025-10-14T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.722176 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.722220 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.722236 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.722259 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.722275 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:49Z","lastTransitionTime":"2025-10-14T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.762217 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:49 crc kubenswrapper[4945]: E1014 15:15:49.762369 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.824122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.824176 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.824187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.824203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.824214 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:49Z","lastTransitionTime":"2025-10-14T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.927527 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.927576 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.927586 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.927603 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:49 crc kubenswrapper[4945]: I1014 15:15:49.927614 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:49Z","lastTransitionTime":"2025-10-14T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.030392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.030424 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.030432 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.030445 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.030453 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:50Z","lastTransitionTime":"2025-10-14T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.133131 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.133195 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.133217 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.133242 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.133498 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:50Z","lastTransitionTime":"2025-10-14T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.242026 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.242276 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.242375 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.242440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.242495 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:50Z","lastTransitionTime":"2025-10-14T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.345906 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.345981 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.346000 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.346025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.346045 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:50Z","lastTransitionTime":"2025-10-14T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.448372 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.449063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.449096 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.449116 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.449128 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:50Z","lastTransitionTime":"2025-10-14T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.552138 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.552202 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.552219 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.552243 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.552261 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:50Z","lastTransitionTime":"2025-10-14T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.655030 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.655099 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.655123 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.655152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.655174 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:50Z","lastTransitionTime":"2025-10-14T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.757481 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.757560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.757577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.757604 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.757622 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:50Z","lastTransitionTime":"2025-10-14T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.762563 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.762580 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.762673 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:50 crc kubenswrapper[4945]: E1014 15:15:50.762813 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:50 crc kubenswrapper[4945]: E1014 15:15:50.762973 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:50 crc kubenswrapper[4945]: E1014 15:15:50.763083 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.777323 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.794273 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb13e41970758fd540b4fff10d1848e4f0da616768856063641a4d1994340c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:39Z\\\",\\\"message\\\":\\\"2025-10-14T15:14:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0\\\\n2025-10-14T15:14:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0 to /host/opt/cni/bin/\\\\n2025-10-14T15:14:54Z [verbose] multus-daemon started\\\\n2025-10-14T15:14:54Z [verbose] Readiness Indicator file check\\\\n2025-10-14T15:15:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.807782 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5fbb433-31b8-4b3c-ae3d-6d34d62cde7f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.820857 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.845602 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:28Z\\\",\\\"message\\\":\\\"openshift-kube-storage-version-migrator-operator for network=default : 1.541893ms\\\\nI1014 15:15:28.613080 6660 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613130 6660 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613145 6660 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1014 15:15:28.613163 6660 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.728948ms\\\\nI1014 15:15:28.612485 6660 obj_re\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.858969 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.861624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.861658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.861667 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.861681 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.861690 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:50Z","lastTransitionTime":"2025-10-14T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.870096 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.880573 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.889742 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.904524 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.919713 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.933588 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.947426 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.959356 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.963548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.963589 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.963601 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.963614 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.963625 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:50Z","lastTransitionTime":"2025-10-14T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.970112 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.984039 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:50 crc kubenswrapper[4945]: I1014 15:15:50.996584 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:50Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.066340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.066382 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.066393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.066409 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.066422 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:51Z","lastTransitionTime":"2025-10-14T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.168278 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.168314 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.168326 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.168342 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.168353 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:51Z","lastTransitionTime":"2025-10-14T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.271544 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.271590 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.271601 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.271618 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.271630 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:51Z","lastTransitionTime":"2025-10-14T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.374303 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.374352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.374370 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.374393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.374413 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:51Z","lastTransitionTime":"2025-10-14T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.477931 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.477970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.477982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.477999 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.478012 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:51Z","lastTransitionTime":"2025-10-14T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.581603 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.581678 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.581702 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.581730 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.581751 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:51Z","lastTransitionTime":"2025-10-14T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.685407 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.685487 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.685512 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.685550 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.685574 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:51Z","lastTransitionTime":"2025-10-14T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.762440 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:51 crc kubenswrapper[4945]: E1014 15:15:51.762626 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.788840 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.788919 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.788933 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.788955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.788993 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:51Z","lastTransitionTime":"2025-10-14T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.892163 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.892219 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.892237 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.892260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.892277 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:51Z","lastTransitionTime":"2025-10-14T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.996341 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.996413 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.996433 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.996463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:51 crc kubenswrapper[4945]: I1014 15:15:51.996482 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:51Z","lastTransitionTime":"2025-10-14T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.100440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.100506 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.100524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.100551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.100570 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:52Z","lastTransitionTime":"2025-10-14T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.202998 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.203074 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.203094 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.203121 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.203138 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:52Z","lastTransitionTime":"2025-10-14T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.305625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.305699 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.305721 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.305749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.305771 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:52Z","lastTransitionTime":"2025-10-14T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.409081 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.409152 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.409176 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.409204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.409226 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:52Z","lastTransitionTime":"2025-10-14T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.512298 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.512366 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.512385 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.512414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.512433 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:52Z","lastTransitionTime":"2025-10-14T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.616419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.616551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.616570 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.616593 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.616611 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:52Z","lastTransitionTime":"2025-10-14T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.720304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.720378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.720397 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.720423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.720440 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:52Z","lastTransitionTime":"2025-10-14T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.762440 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:52 crc kubenswrapper[4945]: E1014 15:15:52.762630 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.762445 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.762753 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:52 crc kubenswrapper[4945]: E1014 15:15:52.762964 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:52 crc kubenswrapper[4945]: E1014 15:15:52.763133 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.822955 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.823011 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.823028 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.823056 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.823074 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:52Z","lastTransitionTime":"2025-10-14T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.926376 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.926418 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.926429 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.926444 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:52 crc kubenswrapper[4945]: I1014 15:15:52.926456 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:52Z","lastTransitionTime":"2025-10-14T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.030329 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.030411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.030436 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.030469 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.030496 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:53Z","lastTransitionTime":"2025-10-14T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.133260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.133312 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.133328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.133352 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.133370 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:53Z","lastTransitionTime":"2025-10-14T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.236180 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.236238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.236255 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.236277 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.236296 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:53Z","lastTransitionTime":"2025-10-14T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.338951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.339022 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.339038 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.339062 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.339080 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:53Z","lastTransitionTime":"2025-10-14T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.441965 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.442016 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.442024 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.442038 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.442046 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:53Z","lastTransitionTime":"2025-10-14T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.545685 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.545768 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.545785 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.545811 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.545828 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:53Z","lastTransitionTime":"2025-10-14T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.648212 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.648249 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.648263 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.648292 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.648313 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:53Z","lastTransitionTime":"2025-10-14T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.752078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.752190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.752215 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.752245 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.752261 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:53Z","lastTransitionTime":"2025-10-14T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.762653 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:53 crc kubenswrapper[4945]: E1014 15:15:53.762862 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.763968 4945 scope.go:117] "RemoveContainer" containerID="99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.854809 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.854915 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.854941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.854973 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.854990 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:53Z","lastTransitionTime":"2025-10-14T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.958431 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.958552 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.958570 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.958596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:53 crc kubenswrapper[4945]: I1014 15:15:53.958613 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:53Z","lastTransitionTime":"2025-10-14T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.063107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.063169 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.063193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.063222 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.063247 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:54Z","lastTransitionTime":"2025-10-14T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.166539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.166587 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.166610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.166639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.166661 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:54Z","lastTransitionTime":"2025-10-14T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.206366 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/2.log" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.209431 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerStarted","Data":"70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134"} Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.209860 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.236535 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.257385 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.269131 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.269206 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.269235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.269266 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.269290 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:54Z","lastTransitionTime":"2025-10-14T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.272984 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.284660 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.303568 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.322178 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.336681 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.351014 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb13e41970758fd540b4fff10d1848e4f0da616768856063641a4d1994340c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:39Z\\\",\\\"message\\\":\\\"2025-10-14T15:14:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0\\\\n2025-10-14T15:14:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0 to /host/opt/cni/bin/\\\\n2025-10-14T15:14:54Z [verbose] multus-daemon started\\\\n2025-10-14T15:14:54Z [verbose] Readiness Indicator file check\\\\n2025-10-14T15:15:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.365392 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5fbb433-31b8-4b3c-ae3d-6d34d62cde7f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.371375 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.371405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.371414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.371428 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.371438 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:54Z","lastTransitionTime":"2025-10-14T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.383767 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.409312 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:28Z\\\",\\\"message\\\":\\\"openshift-kube-storage-version-migrator-operator for network=default : 1.541893ms\\\\nI1014 15:15:28.613080 6660 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613130 6660 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613145 6660 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1014 15:15:28.613163 6660 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.728948ms\\\\nI1014 15:15:28.612485 6660 obj_re\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.423572 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.441171 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.453247 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.463026 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.473264 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.473304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.473315 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.473333 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.473345 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:54Z","lastTransitionTime":"2025-10-14T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.477268 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.490229 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:54Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.574869 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.574927 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.574937 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.574951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.574960 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:54Z","lastTransitionTime":"2025-10-14T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.583431 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.583571 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:16:58.583549359 +0000 UTC m=+148.567597727 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.583613 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.583685 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.583778 4945 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.583809 4945 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.583868 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:16:58.583845827 +0000 UTC m=+148.567894225 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.583928 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 15:16:58.583920019 +0000 UTC m=+148.567968387 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.677259 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.677304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.677316 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.677332 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.677345 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:54Z","lastTransitionTime":"2025-10-14T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.684786 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.684841 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.684956 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.684977 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.684987 4945 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.685025 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 15:16:58.685011568 +0000 UTC m=+148.669059936 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.684956 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.685084 4945 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.685099 4945 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.685142 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 15:16:58.685130681 +0000 UTC m=+148.669179049 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.762735 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.762813 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.762738 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.762897 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.763004 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:54 crc kubenswrapper[4945]: E1014 15:15:54.763116 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.780146 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.780192 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.780203 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.780221 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.780234 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:54Z","lastTransitionTime":"2025-10-14T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.882782 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.882834 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.882847 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.882864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.882901 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:54Z","lastTransitionTime":"2025-10-14T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.984664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.984709 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.984719 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.984734 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:54 crc kubenswrapper[4945]: I1014 15:15:54.984745 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:54Z","lastTransitionTime":"2025-10-14T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.087516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.087558 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.087570 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.087611 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.087622 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:55Z","lastTransitionTime":"2025-10-14T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.189852 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.189924 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.189941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.189963 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.189980 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:55Z","lastTransitionTime":"2025-10-14T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.291851 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.291977 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.292003 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.292036 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.292057 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:55Z","lastTransitionTime":"2025-10-14T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.394829 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.394926 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.394944 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.394970 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.394988 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:55Z","lastTransitionTime":"2025-10-14T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.500047 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.500108 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.500124 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.500156 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.500173 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:55Z","lastTransitionTime":"2025-10-14T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.603054 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.603085 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.603095 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.603110 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.603121 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:55Z","lastTransitionTime":"2025-10-14T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.707160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.707426 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.707440 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.707458 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.707470 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:55Z","lastTransitionTime":"2025-10-14T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.762239 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:55 crc kubenswrapper[4945]: E1014 15:15:55.762480 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.810408 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.810511 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.810535 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.810561 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.810579 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:55Z","lastTransitionTime":"2025-10-14T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.914028 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.914081 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.914100 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.914126 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:55 crc kubenswrapper[4945]: I1014 15:15:55.914144 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:55Z","lastTransitionTime":"2025-10-14T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.017349 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.017407 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.017427 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.017450 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.017467 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:56Z","lastTransitionTime":"2025-10-14T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.119639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.119701 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.119721 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.119745 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.119764 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:56Z","lastTransitionTime":"2025-10-14T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.220970 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/3.log" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.221989 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/2.log" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.222245 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.222340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.222361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.222420 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.222440 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:56Z","lastTransitionTime":"2025-10-14T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.226463 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134" exitCode=1 Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.226529 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134"} Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.226589 4945 scope.go:117] "RemoveContainer" containerID="99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.228668 4945 scope.go:117] "RemoveContainer" containerID="70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134" Oct 14 15:15:56 crc kubenswrapper[4945]: E1014 15:15:56.229100 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.254462 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.277028 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.291275 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.305510 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.321371 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.325569 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.325615 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.325624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.325640 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.325650 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:56Z","lastTransitionTime":"2025-10-14T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.341578 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.358671 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.376319 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.389653 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.401384 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.414504 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.427559 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.427593 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.427609 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.427625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.427637 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:56Z","lastTransitionTime":"2025-10-14T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.430548 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5fbb433-31b8-4b3c-ae3d-6d34d62cde7f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.443860 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.462242 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb13e41970758fd540b4fff10d1848e4f0da616768856063641a4d1994340c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:39Z\\\",\\\"message\\\":\\\"2025-10-14T15:14:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0\\\\n2025-10-14T15:14:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0 to /host/opt/cni/bin/\\\\n2025-10-14T15:14:54Z [verbose] multus-daemon started\\\\n2025-10-14T15:14:54Z [verbose] Readiness Indicator file check\\\\n2025-10-14T15:15:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.479105 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.488115 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.516878 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:28Z\\\",\\\"message\\\":\\\"openshift-kube-storage-version-migrator-operator for network=default : 1.541893ms\\\\nI1014 15:15:28.613080 6660 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613130 6660 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613145 6660 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1014 15:15:28.613163 6660 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.728948ms\\\\nI1014 15:15:28.612485 6660 obj_re\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:55Z\\\",\\\"message\\\":\\\" 1 for removal\\\\nI1014 15:15:54.844390 7028 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 15:15:54.844330 7028 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1014 15:15:54.844413 7028 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 15:15:54.844430 7028 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 15:15:54.844484 7028 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 15:15:54.844332 7028 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 15:15:54.844519 7028 factory.go:656] Stopping watch factory\\\\nI1014 15:15:54.844528 7028 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 15:15:54.844548 7028 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 15:15:54.844572 7028 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 15:15:54.844588 7028 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 15:15:54.844611 7028 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 15:15:54.844627 7028 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:56Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.531240 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.531279 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.531295 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.531318 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.531336 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:56Z","lastTransitionTime":"2025-10-14T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.633628 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.634234 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.634362 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.634527 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.634670 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:56Z","lastTransitionTime":"2025-10-14T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.737992 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.738339 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.738478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.738655 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.738808 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:56Z","lastTransitionTime":"2025-10-14T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.762684 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.762833 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.763116 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:56 crc kubenswrapper[4945]: E1014 15:15:56.763356 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:56 crc kubenswrapper[4945]: E1014 15:15:56.763497 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:56 crc kubenswrapper[4945]: E1014 15:15:56.763603 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.780721 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.842860 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.843372 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.843663 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.843957 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.844224 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:56Z","lastTransitionTime":"2025-10-14T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.948349 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.948655 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.949162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.949368 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:56 crc kubenswrapper[4945]: I1014 15:15:56.949728 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:56Z","lastTransitionTime":"2025-10-14T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.053419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.053516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.053534 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.054025 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.054084 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:57Z","lastTransitionTime":"2025-10-14T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.157009 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.157121 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.157145 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.157222 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.157253 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:57Z","lastTransitionTime":"2025-10-14T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.233433 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/3.log" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.259655 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.259722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.259741 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.259763 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.259781 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:57Z","lastTransitionTime":"2025-10-14T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.362109 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.362168 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.362191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.362221 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.362243 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:57Z","lastTransitionTime":"2025-10-14T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.465407 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.465471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.465493 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.465517 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.465533 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:57Z","lastTransitionTime":"2025-10-14T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.569266 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.569333 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.569349 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.569374 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.569390 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:57Z","lastTransitionTime":"2025-10-14T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.672102 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.672138 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.672148 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.672162 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.672171 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:57Z","lastTransitionTime":"2025-10-14T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.761837 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:57 crc kubenswrapper[4945]: E1014 15:15:57.762021 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.775861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.775922 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.775936 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.775954 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.775966 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:57Z","lastTransitionTime":"2025-10-14T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.879474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.879560 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.879581 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.879607 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.879625 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:57Z","lastTransitionTime":"2025-10-14T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.982965 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.983005 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.983013 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.983028 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:57 crc kubenswrapper[4945]: I1014 15:15:57.983036 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:57Z","lastTransitionTime":"2025-10-14T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.085337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.085369 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.085378 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.085391 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.085401 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.188328 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.188397 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.188417 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.188443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.188462 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.290610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.290653 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.290665 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.290680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.290689 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.393265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.393323 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.393340 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.393363 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.393381 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.404520 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.404639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.404658 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.404680 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.404695 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: E1014 15:15:58.424984 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.430780 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.430857 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.430921 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.430974 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.431000 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: E1014 15:15:58.451940 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.457605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.457670 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.457686 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.457712 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.457734 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: E1014 15:15:58.478397 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.483346 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.483442 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.483569 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.483598 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.483618 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: E1014 15:15:58.501730 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.506626 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.506699 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.506730 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.506762 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.506783 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: E1014 15:15:58.523593 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:15:58Z is after 2025-08-24T17:21:41Z" Oct 14 15:15:58 crc kubenswrapper[4945]: E1014 15:15:58.523854 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.526142 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.526249 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.526271 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.526301 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.526321 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.628785 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.628859 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.628886 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.628901 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.628916 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.737486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.737574 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.737599 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.737630 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.737657 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.762399 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.762420 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:15:58 crc kubenswrapper[4945]: E1014 15:15:58.762675 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.762765 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:15:58 crc kubenswrapper[4945]: E1014 15:15:58.762795 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:15:58 crc kubenswrapper[4945]: E1014 15:15:58.762977 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.840297 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.840359 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.840383 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.840405 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.840426 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.943549 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.943607 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.943625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.943653 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:58 crc kubenswrapper[4945]: I1014 15:15:58.943670 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:58Z","lastTransitionTime":"2025-10-14T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.046495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.046551 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.046568 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.046587 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.046600 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:59Z","lastTransitionTime":"2025-10-14T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.149610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.149675 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.149692 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.149716 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.149733 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:59Z","lastTransitionTime":"2025-10-14T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.252183 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.252248 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.252265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.252290 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.252308 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:59Z","lastTransitionTime":"2025-10-14T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.355261 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.355318 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.355327 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.355342 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.355375 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:59Z","lastTransitionTime":"2025-10-14T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.457583 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.457635 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.457656 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.457679 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.457696 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:59Z","lastTransitionTime":"2025-10-14T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.560464 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.560542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.560568 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.560600 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.560622 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:59Z","lastTransitionTime":"2025-10-14T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.663613 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.663683 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.663699 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.663722 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.663740 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:59Z","lastTransitionTime":"2025-10-14T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.762836 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:15:59 crc kubenswrapper[4945]: E1014 15:15:59.763740 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.766155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.766280 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.766302 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.766329 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.766346 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:59Z","lastTransitionTime":"2025-10-14T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.869107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.869298 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.869323 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.869351 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.869367 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:59Z","lastTransitionTime":"2025-10-14T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.972640 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.972770 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.972794 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.972824 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:15:59 crc kubenswrapper[4945]: I1014 15:15:59.972847 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:15:59Z","lastTransitionTime":"2025-10-14T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.076379 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.076487 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.076512 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.076598 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.076710 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:00Z","lastTransitionTime":"2025-10-14T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.179690 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.179749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.179771 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.179838 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.180124 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:00Z","lastTransitionTime":"2025-10-14T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.283020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.283061 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.283073 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.283090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.283102 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:00Z","lastTransitionTime":"2025-10-14T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.385445 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.385497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.385511 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.385533 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.385548 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:00Z","lastTransitionTime":"2025-10-14T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.488225 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.488262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.488270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.488284 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.488293 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:00Z","lastTransitionTime":"2025-10-14T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.603167 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.603232 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.603250 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.603274 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.603290 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:00Z","lastTransitionTime":"2025-10-14T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.706145 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.706190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.706205 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.706245 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.706259 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:00Z","lastTransitionTime":"2025-10-14T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.762193 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:00 crc kubenswrapper[4945]: E1014 15:16:00.762443 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.762485 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.762566 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:00 crc kubenswrapper[4945]: E1014 15:16:00.762722 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:00 crc kubenswrapper[4945]: E1014 15:16:00.762864 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.776616 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea50d95-f180-4c5a-b9f5-0ca1a99cfa08\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbbd5b1a7af200f7b8fe75b4235faf670536f59af96ed16e739611b5243ee6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6be28c1218c1a6d3076c3cb17faac5f45232bc665d9590e92e2ae3319775f14d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6be28c1218c1a6d3076c3cb17faac5f45232bc665d9590e92e2ae3319775f14d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.794422 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.808918 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.808978 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.809000 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.809030 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.809052 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:00Z","lastTransitionTime":"2025-10-14T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.811723 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.834935 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99ff1f1a6070c0632ae1cf3bc00d4cca18b78e6b072f8bc7d313a32b79f2dd03\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:28Z\\\",\\\"message\\\":\\\"openshift-kube-storage-version-migrator-operator for network=default : 1.541893ms\\\\nI1014 15:15:28.613080 6660 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613130 6660 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 15:15:28.613145 6660 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI1014 15:15:28.613163 6660 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.728948ms\\\\nI1014 15:15:28.612485 6660 obj_re\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:55Z\\\",\\\"message\\\":\\\" 1 for removal\\\\nI1014 15:15:54.844390 7028 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 15:15:54.844330 7028 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1014 15:15:54.844413 7028 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 15:15:54.844430 7028 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 15:15:54.844484 7028 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 15:15:54.844332 7028 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 15:15:54.844519 7028 factory.go:656] Stopping watch factory\\\\nI1014 15:15:54.844528 7028 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 15:15:54.844548 7028 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 15:15:54.844572 7028 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 15:15:54.844588 7028 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 15:15:54.844611 7028 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 15:15:54.844627 7028 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.860162 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.879243 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.890836 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.904230 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.911571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.911630 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.911649 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.911675 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.911694 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:00Z","lastTransitionTime":"2025-10-14T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.916710 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.929746 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.943217 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.961117 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.973949 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:00 crc kubenswrapper[4945]: I1014 15:16:00.988863 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:00Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.003328 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:01Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.014165 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.014217 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.014235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.014258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.014274 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:01Z","lastTransitionTime":"2025-10-14T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.018556 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5fbb433-31b8-4b3c-ae3d-6d34d62cde7f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:01Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.034805 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:01Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.054318 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb13e41970758fd540b4fff10d1848e4f0da616768856063641a4d1994340c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:39Z\\\",\\\"message\\\":\\\"2025-10-14T15:14:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0\\\\n2025-10-14T15:14:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0 to /host/opt/cni/bin/\\\\n2025-10-14T15:14:54Z [verbose] multus-daemon started\\\\n2025-10-14T15:14:54Z [verbose] Readiness Indicator file check\\\\n2025-10-14T15:15:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:01Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.117646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.117701 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.117718 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.117740 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.117758 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:01Z","lastTransitionTime":"2025-10-14T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.220230 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.220291 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.220308 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.220332 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.220349 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:01Z","lastTransitionTime":"2025-10-14T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.323109 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.323172 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.323187 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.323211 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.323227 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:01Z","lastTransitionTime":"2025-10-14T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.425836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.425941 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.425966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.425995 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.426016 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:01Z","lastTransitionTime":"2025-10-14T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.528810 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.528899 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.528913 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.528930 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.528942 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:01Z","lastTransitionTime":"2025-10-14T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.632356 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.632519 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.632541 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.632564 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.632583 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:01Z","lastTransitionTime":"2025-10-14T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.735305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.735343 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.735355 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.735372 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.735383 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:01Z","lastTransitionTime":"2025-10-14T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.762658 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:01 crc kubenswrapper[4945]: E1014 15:16:01.762941 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.838194 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.838226 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.838238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.838254 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.838265 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:01Z","lastTransitionTime":"2025-10-14T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.940775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.940827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.940841 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.940861 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:01 crc kubenswrapper[4945]: I1014 15:16:01.940893 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:01Z","lastTransitionTime":"2025-10-14T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.043293 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.043333 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.043345 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.043361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.043373 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:02Z","lastTransitionTime":"2025-10-14T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.145496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.145521 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.145529 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.145542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.145550 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:02Z","lastTransitionTime":"2025-10-14T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.248490 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.248531 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.248542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.248558 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.248569 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:02Z","lastTransitionTime":"2025-10-14T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.351154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.351225 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.351244 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.351269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.351287 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:02Z","lastTransitionTime":"2025-10-14T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.453755 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.453827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.453869 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.453947 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.453972 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:02Z","lastTransitionTime":"2025-10-14T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.556636 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.556717 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.556750 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.556793 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.556951 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:02Z","lastTransitionTime":"2025-10-14T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.660716 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.660775 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.660814 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.660849 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.660872 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:02Z","lastTransitionTime":"2025-10-14T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.762240 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:02 crc kubenswrapper[4945]: E1014 15:16:02.762383 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.763015 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:02 crc kubenswrapper[4945]: E1014 15:16:02.763090 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.763244 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:02 crc kubenswrapper[4945]: E1014 15:16:02.763489 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.763783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.763978 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.763997 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.764016 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.764031 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:02Z","lastTransitionTime":"2025-10-14T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.866133 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.866177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.866191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.866206 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.866218 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:02Z","lastTransitionTime":"2025-10-14T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.969470 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.969516 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.969525 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.969540 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:02 crc kubenswrapper[4945]: I1014 15:16:02.969552 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:02Z","lastTransitionTime":"2025-10-14T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.072393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.072439 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.072457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.072479 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.072493 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:03Z","lastTransitionTime":"2025-10-14T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.174501 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.174548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.174576 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.174591 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.174601 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:03Z","lastTransitionTime":"2025-10-14T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.276553 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.276590 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.276601 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.276620 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.276634 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:03Z","lastTransitionTime":"2025-10-14T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.379327 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.379383 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.379400 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.379424 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.379445 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:03Z","lastTransitionTime":"2025-10-14T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.482784 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.482857 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.482925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.482956 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.482979 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:03Z","lastTransitionTime":"2025-10-14T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.585820 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.585917 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.585936 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.585967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.585985 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:03Z","lastTransitionTime":"2025-10-14T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.688743 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.688811 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.688834 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.688863 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.688922 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:03Z","lastTransitionTime":"2025-10-14T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.762376 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:03 crc kubenswrapper[4945]: E1014 15:16:03.762606 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.793143 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.793226 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.793246 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.793277 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.793301 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:03Z","lastTransitionTime":"2025-10-14T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.896245 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.896298 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.896314 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.896336 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:03 crc kubenswrapper[4945]: I1014 15:16:03.896352 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:03Z","lastTransitionTime":"2025-10-14T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:03.999969 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.000066 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.000085 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.000122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.000142 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:04Z","lastTransitionTime":"2025-10-14T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.103674 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.103754 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.103777 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.103806 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.103831 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:04Z","lastTransitionTime":"2025-10-14T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.207040 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.207111 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.207131 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.207154 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.207170 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:04Z","lastTransitionTime":"2025-10-14T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.310248 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.310318 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.310337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.310360 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.310380 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:04Z","lastTransitionTime":"2025-10-14T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.413419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.413497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.413536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.413571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.413597 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:04Z","lastTransitionTime":"2025-10-14T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.517333 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.517409 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.517605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.517636 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.517652 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:04Z","lastTransitionTime":"2025-10-14T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.622653 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.622710 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.622723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.622742 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.622753 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:04Z","lastTransitionTime":"2025-10-14T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.725382 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.725455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.725466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.725486 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.725499 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:04Z","lastTransitionTime":"2025-10-14T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.762709 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.762758 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:04 crc kubenswrapper[4945]: E1014 15:16:04.762846 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.762856 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:04 crc kubenswrapper[4945]: E1014 15:16:04.763030 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:04 crc kubenswrapper[4945]: E1014 15:16:04.763183 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.828991 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.829065 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.829088 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.829118 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.829139 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:04Z","lastTransitionTime":"2025-10-14T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.936925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.936993 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.937012 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.937039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:04 crc kubenswrapper[4945]: I1014 15:16:04.937059 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:04Z","lastTransitionTime":"2025-10-14T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.039402 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.039441 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.039449 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.039463 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.039473 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:05Z","lastTransitionTime":"2025-10-14T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.150499 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.150538 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.150547 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.150561 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.150570 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:05Z","lastTransitionTime":"2025-10-14T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.252946 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.252989 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.253001 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.253023 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.253036 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:05Z","lastTransitionTime":"2025-10-14T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.356807 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.356920 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.356954 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.356991 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.357014 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:05Z","lastTransitionTime":"2025-10-14T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.460566 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.460611 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.460622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.460640 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.460651 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:05Z","lastTransitionTime":"2025-10-14T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.564010 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.564078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.564098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.564125 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.564144 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:05Z","lastTransitionTime":"2025-10-14T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.667307 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.667357 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.667372 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.667398 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.667414 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:05Z","lastTransitionTime":"2025-10-14T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.762131 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:05 crc kubenswrapper[4945]: E1014 15:16:05.762336 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.770711 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.770766 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.770778 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.770799 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.770813 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:05Z","lastTransitionTime":"2025-10-14T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.874686 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.875091 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.875304 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.875571 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.875826 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:05Z","lastTransitionTime":"2025-10-14T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.980606 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.981077 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.981122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.981160 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:05 crc kubenswrapper[4945]: I1014 15:16:05.981183 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:05Z","lastTransitionTime":"2025-10-14T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.084565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.084622 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.084640 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.084664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.084681 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:06Z","lastTransitionTime":"2025-10-14T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.186763 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.187229 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.187248 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.187275 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.187292 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:06Z","lastTransitionTime":"2025-10-14T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.290219 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.290269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.290287 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.290307 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.290318 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:06Z","lastTransitionTime":"2025-10-14T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.392894 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.392948 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.392960 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.392979 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.392992 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:06Z","lastTransitionTime":"2025-10-14T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.495771 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.495847 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.495860 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.495908 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.495927 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:06Z","lastTransitionTime":"2025-10-14T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.598327 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.598399 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.598412 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.598443 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.598456 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:06Z","lastTransitionTime":"2025-10-14T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.702158 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.702208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.702221 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.702254 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.702271 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:06Z","lastTransitionTime":"2025-10-14T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.762571 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:06 crc kubenswrapper[4945]: E1014 15:16:06.762820 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.763120 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.763151 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:06 crc kubenswrapper[4945]: E1014 15:16:06.763285 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:06 crc kubenswrapper[4945]: E1014 15:16:06.763454 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.805191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.805252 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.805274 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.805301 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.805317 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:06Z","lastTransitionTime":"2025-10-14T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.909269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.909358 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.909385 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.909423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:06 crc kubenswrapper[4945]: I1014 15:16:06.909450 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:06Z","lastTransitionTime":"2025-10-14T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.012467 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.012534 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.012549 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.012574 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.012596 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:07Z","lastTransitionTime":"2025-10-14T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.115948 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.116211 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.116235 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.116261 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.116279 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:07Z","lastTransitionTime":"2025-10-14T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.220596 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.220674 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.220699 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.220741 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.220770 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:07Z","lastTransitionTime":"2025-10-14T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.324330 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.324411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.324435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.324471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.324497 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:07Z","lastTransitionTime":"2025-10-14T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.427967 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.428053 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.428079 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.428113 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.428133 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:07Z","lastTransitionTime":"2025-10-14T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.532086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.532159 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.532171 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.532191 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.532202 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:07Z","lastTransitionTime":"2025-10-14T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.635535 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.635603 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.635616 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.635637 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.635653 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:07Z","lastTransitionTime":"2025-10-14T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.739123 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.739161 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.739173 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.739217 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.739229 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:07Z","lastTransitionTime":"2025-10-14T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.762077 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:07 crc kubenswrapper[4945]: E1014 15:16:07.762275 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.844010 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.844059 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.844069 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.844086 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.844097 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:07Z","lastTransitionTime":"2025-10-14T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.946729 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.946762 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.946770 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.946786 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:07 crc kubenswrapper[4945]: I1014 15:16:07.946814 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:07Z","lastTransitionTime":"2025-10-14T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.049175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.049223 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.049233 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.049250 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.049262 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.152431 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.152471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.152480 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.152497 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.152506 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.255830 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.255940 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.255964 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.255995 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.256019 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.358870 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.358987 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.359015 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.359048 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.359071 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.462129 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.462174 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.462188 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.462210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.462225 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.564835 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.564896 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.564908 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.564925 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.564936 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.667627 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.667696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.667713 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.667736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.667754 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.762783 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.762956 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:08 crc kubenswrapper[4945]: E1014 15:16:08.763100 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.763135 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:08 crc kubenswrapper[4945]: E1014 15:16:08.763328 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:08 crc kubenswrapper[4945]: E1014 15:16:08.763791 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.764309 4945 scope.go:117] "RemoveContainer" containerID="70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134" Oct 14 15:16:08 crc kubenswrapper[4945]: E1014 15:16:08.764553 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.771669 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.771768 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.771784 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.771812 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.771829 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.784184 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5fbb433-31b8-4b3c-ae3d-6d34d62cde7f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad138328a095b9f6820c121d88bc27f50ada4a2fd2dbeac0dda766e798ff60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cf4b43b3382b3d2b547be2f42b1dd4571965a7ef4c019215ba0ea32bae920b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4aa6a4c325f071791dbd7ed83cc586bfccce3ace8145af67582cf651851ed467\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dac6be0967fdeda2965eb5fad2e9e1489f2821c3436f38b7c1ecc97ec09b9fa2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.797635 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.817952 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b76xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82b9cd9-7a14-4dd9-bb0e-fde942389666\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb13e41970758fd540b4fff10d1848e4f0da616768856063641a4d1994340c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:39Z\\\",\\\"message\\\":\\\"2025-10-14T15:14:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0\\\\n2025-10-14T15:14:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7334cc24-5b3f-44cd-86d0-80042feb7cb0 to /host/opt/cni/bin/\\\\n2025-10-14T15:14:54Z [verbose] multus-daemon started\\\\n2025-10-14T15:14:54Z [verbose] Readiness Indicator file check\\\\n2025-10-14T15:15:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b76xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.837282 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.849475 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mgprj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3c99b5a-8a60-4cb6-8072-47c4a78d3d4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dd3642a886d4335e626dc1951831e8bf445d6d8f426e464099b91d77059a106\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g59cr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mgprj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.874940 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.875001 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.875020 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.875047 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.875067 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.880698 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6de8ddf6-38db-4394-8660-7e9659bf2bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T15:15:55Z\\\",\\\"message\\\":\\\" 1 for removal\\\\nI1014 15:15:54.844390 7028 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 15:15:54.844330 7028 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1014 15:15:54.844413 7028 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 15:15:54.844430 7028 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 15:15:54.844484 7028 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 15:15:54.844332 7028 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 15:15:54.844519 7028 factory.go:656] Stopping watch factory\\\\nI1014 15:15:54.844528 7028 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 15:15:54.844548 7028 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 15:15:54.844572 7028 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 15:15:54.844588 7028 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 15:15:54.844611 7028 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 15:15:54.844627 7028 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:15:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62tc9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pqtt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.885540 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.885592 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.885610 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.885663 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.885686 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.893666 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea50d95-f180-4c5a-b9f5-0ca1a99cfa08\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbbd5b1a7af200f7b8fe75b4235faf670536f59af96ed16e739611b5243ee6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6be28c1218c1a6d3076c3cb17faac5f45232bc665d9590e92e2ae3319775f14d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6be28c1218c1a6d3076c3cb17faac5f45232bc665d9590e92e2ae3319775f14d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: E1014 15:16:08.902077 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.907783 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.907809 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.907819 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.907836 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.907848 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.914270 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ecfd12-913b-4802-99cc-11dfb3d1a0f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34e74f28d33eef348d7781084938a784fa29d5621617f9302e88e2629a22f028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30a1c09f07f8113058c3e949256e2701a3212b963158ad32e979a849e7fce74b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66da3b45a68cf56d2248ede70ea741b1db4cdc88c00216ace53380015eac3abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd461ce71a00f52fd22c20a3613a47f8a8e0e70f113a15b9759330e574a4e046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f58850898a6df17e1b17e1636208b00be4b375bc59114e990374a2cb8c59b15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e91718cc9409e64ba6a9758d5a43b9815e783182aa6e86b34052cea9f89d7c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06b90c156d35054cc7bb8e47fc8952df3940b9e9e059ae4c767586e0fcf1e107\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nhpg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lnbkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: E1014 15:16:08.926208 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.927663 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08049862-1ea7-4efc-b9ed-df61b47d430b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd52a9e67c479329e454b360491af7b62e0133a8d0e8384084aca4e2985b317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qch55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tx7tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.929664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.929695 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.929709 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.929728 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.929740 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.948983 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a336c34-8669-4fbe-8f51-c339403782b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62f545cf12c0e871f67c002369d037de2c50bc45af48ef03781876e837dfd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d6693bdaf8e53b11d93efc5b1d039e3a78393b28c267b0717ea272b0c5be59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk2nw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmsrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: E1014 15:16:08.952917 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.957144 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.957185 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.957197 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.957216 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.957267 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.959690 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be17bc84-909e-4150-8ee7-9a7d23173ba0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jvc8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:15:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ftg5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: E1014 15:16:08.970867 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.975417 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.975464 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.975474 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.975495 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.975508 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.975992 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e296950-8549-4151-a51c-5a9481f3407a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:15:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae23c97686e3f5f48749071e72b87adb505e62172bb12b721ede6c17b839b83d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973e1519c3cb8ea706abea22a85124bf291b69e487cc0450e480f13d48d828bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://131aeaddb9c25e673b7ec5bc367ad083f1a62965b186945024d7265758fa65ce\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd8c401986732d7c0931bd47856c7f13b6058947d4d19ad611f3210dfb247f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b357e31f9908afb7f9f423e63a41d449251aad1fb4d9873bb8b8524ec42b98\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T15:14:44Z\\\",\\\"message\\\":\\\"W1014 15:14:33.963210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 15:14:33.964142 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760454873 cert, and key in /tmp/serving-cert-1001518689/serving-signer.crt, /tmp/serving-cert-1001518689/serving-signer.key\\\\nI1014 15:14:34.230174 1 observer_polling.go:159] Starting file observer\\\\nW1014 15:14:34.234657 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 15:14:34.235018 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 15:14:34.236367 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1001518689/tls.crt::/tmp/serving-cert-1001518689/tls.key\\\\\\\"\\\\nF1014 15:14:44.899897 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://333d13ffbe4cfdb6a0ca062362c9920994e25b00d23b3683c2526db772c188d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cd2bb0952191a7be5bb01a348dbb607db7055d63d6c39f0f53bb8716831f4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T15:14:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: E1014 15:16:08.988569 4945 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T15:16:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d936a6ed-46ff-4fc3-bc0a-62bc04b87659\\\",\\\"systemUUID\\\":\\\"6a76cd5d-bafd-45c2-978d-e5848e35a4cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: E1014 15:16:08.989028 4945 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.990171 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e3b9352d0a471afa78e605455e44e0359a9e9083779030917f8178c3373cd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:08Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.991478 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.991625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.991735 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.992088 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:08 crc kubenswrapper[4945]: I1014 15:16:08.992220 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:08Z","lastTransitionTime":"2025-10-14T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.047641 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://245e540c07dfafec6d80d94ce731c3f0663a610ffc3b23d6b5a312f29908a46f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8d0efe69003e305f5f683c9896628f1e63d4da1c111aca0e4ded917c020efd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:09Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.060933 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:09Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.072608 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca5b415e124c1ea967bf67189dcb25a60cba496ea484cd5fd442ed2c53416a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:09Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.083431 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zfc4b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7648f858-37c9-4f14-9d0f-5e3b354f4e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d8613e5d3c32c51b2a07df51fa3bf1b9fcebf8797a08bf05ec0710a315c05ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6snbc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zfc4b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:09Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.095223 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.095252 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.095260 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.095274 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.095284 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:09Z","lastTransitionTime":"2025-10-14T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.096337 4945 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e0d4108-a152-46b2-96ba-3b0a710acf22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T15:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3d42d6db5950df618ed0bbf52434f003e918121b768260e954c20ac8596fd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://509035a23774569851494d20852cf95f06deb3186ba092f3821e4e7f35432e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8d7a98908ca8d98ff0f7c2c4824e2f13f38190bcab78d89afe43755cd6551b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3642723ae863fa1ae789d075b414a04e78ed378b98c3521a8a9ed3e9e91d61a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T15:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T15:14:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T15:16:09Z is after 2025-08-24T17:21:41Z" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.198821 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.199204 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.199397 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.199598 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.199795 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:09Z","lastTransitionTime":"2025-10-14T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.305696 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.305747 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.305764 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.305788 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.305804 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:09Z","lastTransitionTime":"2025-10-14T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.409583 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.409652 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.409663 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.409686 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.409698 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:09Z","lastTransitionTime":"2025-10-14T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.514256 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.514668 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.514797 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.515063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.515284 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:09Z","lastTransitionTime":"2025-10-14T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.619177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.619542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.619726 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.619922 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.620107 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:09Z","lastTransitionTime":"2025-10-14T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.723157 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.723636 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.723828 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.724168 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.724362 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:09Z","lastTransitionTime":"2025-10-14T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.762448 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:09 crc kubenswrapper[4945]: E1014 15:16:09.763016 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.827349 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.827434 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.827457 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.827530 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.827556 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:09Z","lastTransitionTime":"2025-10-14T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.930621 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.930663 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.930676 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.930692 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.930706 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:09Z","lastTransitionTime":"2025-10-14T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:09 crc kubenswrapper[4945]: I1014 15:16:09.957519 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:09 crc kubenswrapper[4945]: E1014 15:16:09.957691 4945 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:16:09 crc kubenswrapper[4945]: E1014 15:16:09.957777 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs podName:be17bc84-909e-4150-8ee7-9a7d23173ba0 nodeName:}" failed. No retries permitted until 2025-10-14 15:17:13.95775535 +0000 UTC m=+163.941803718 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs") pod "network-metrics-daemon-ftg5h" (UID: "be17bc84-909e-4150-8ee7-9a7d23173ba0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.034013 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.034063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.034080 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.034106 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.034121 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:10Z","lastTransitionTime":"2025-10-14T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.137072 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.137132 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.137151 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.137177 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.137194 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:10Z","lastTransitionTime":"2025-10-14T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.239653 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.239683 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.239691 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.239730 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.239742 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:10Z","lastTransitionTime":"2025-10-14T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.342857 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.342947 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.342966 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.342994 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.343011 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:10Z","lastTransitionTime":"2025-10-14T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.445945 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.445982 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.445991 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.446003 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.446012 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:10Z","lastTransitionTime":"2025-10-14T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.548532 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.548565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.548574 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.548587 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.548598 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:10Z","lastTransitionTime":"2025-10-14T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.650980 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.651041 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.651058 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.651081 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.651099 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:10Z","lastTransitionTime":"2025-10-14T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.754210 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.754269 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.754287 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.754322 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.754356 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:10Z","lastTransitionTime":"2025-10-14T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.762090 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:10 crc kubenswrapper[4945]: E1014 15:16:10.762215 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.762100 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:10 crc kubenswrapper[4945]: E1014 15:16:10.762586 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.762575 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:10 crc kubenswrapper[4945]: E1014 15:16:10.765869 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.845194 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=46.84515229 podStartE2EDuration="46.84515229s" podCreationTimestamp="2025-10-14 15:15:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:10.842862696 +0000 UTC m=+100.826911064" watchObservedRunningTime="2025-10-14 15:16:10.84515229 +0000 UTC m=+100.829200668" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.845437 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-b76xs" podStartSLOduration=79.845430108 podStartE2EDuration="1m19.845430108s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:10.827449723 +0000 UTC m=+100.811498101" watchObservedRunningTime="2025-10-14 15:16:10.845430108 +0000 UTC m=+100.829478486" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.857595 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.857634 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.857643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.857662 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.857673 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:10Z","lastTransitionTime":"2025-10-14T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.903028 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-mgprj" podStartSLOduration=80.902995785 podStartE2EDuration="1m20.902995785s" podCreationTimestamp="2025-10-14 15:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:10.861014436 +0000 UTC m=+100.845062804" watchObservedRunningTime="2025-10-14 15:16:10.902995785 +0000 UTC m=+100.887044163" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.952076 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=14.952041652 podStartE2EDuration="14.952041652s" podCreationTimestamp="2025-10-14 15:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:10.930463106 +0000 UTC m=+100.914511474" watchObservedRunningTime="2025-10-14 15:16:10.952041652 +0000 UTC m=+100.936090030" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.960392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.960442 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.960460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.960503 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.960521 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:10Z","lastTransitionTime":"2025-10-14T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.983613 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podStartSLOduration=79.983588538 podStartE2EDuration="1m19.983588538s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:10.968057312 +0000 UTC m=+100.952105700" watchObservedRunningTime="2025-10-14 15:16:10.983588538 +0000 UTC m=+100.967636896" Oct 14 15:16:10 crc kubenswrapper[4945]: I1014 15:16:10.984236 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmsrd" podStartSLOduration=79.984204425 podStartE2EDuration="1m19.984204425s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:10.983106795 +0000 UTC m=+100.967155163" watchObservedRunningTime="2025-10-14 15:16:10.984204425 +0000 UTC m=+100.968252813" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.013207 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=81.013185169 podStartE2EDuration="1m21.013185169s" podCreationTimestamp="2025-10-14 15:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:11.012399067 +0000 UTC m=+100.996447445" watchObservedRunningTime="2025-10-14 15:16:11.013185169 +0000 UTC m=+100.997233537" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.049620 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-lnbkr" podStartSLOduration=80.049596722 podStartE2EDuration="1m20.049596722s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:11.034554529 +0000 UTC m=+101.018602897" watchObservedRunningTime="2025-10-14 15:16:11.049596722 +0000 UTC m=+101.033645090" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.063396 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.063460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.063485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.063505 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.063520 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:11Z","lastTransitionTime":"2025-10-14T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.114082 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.114056682 podStartE2EDuration="1m18.114056682s" podCreationTimestamp="2025-10-14 15:14:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:11.113059864 +0000 UTC m=+101.097108252" watchObservedRunningTime="2025-10-14 15:16:11.114056682 +0000 UTC m=+101.098105050" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.114658 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-zfc4b" podStartSLOduration=81.114652679 podStartE2EDuration="1m21.114652679s" podCreationTimestamp="2025-10-14 15:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:11.090098469 +0000 UTC m=+101.074146847" watchObservedRunningTime="2025-10-14 15:16:11.114652679 +0000 UTC m=+101.098701047" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.166183 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.166239 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.166253 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.166275 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.166287 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:11Z","lastTransitionTime":"2025-10-14T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.269575 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.269643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.269656 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.269679 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.269699 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:11Z","lastTransitionTime":"2025-10-14T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.372719 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.372782 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.372795 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.372816 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.372834 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:11Z","lastTransitionTime":"2025-10-14T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.475772 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.475814 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.475823 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.475840 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.475851 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:11Z","lastTransitionTime":"2025-10-14T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.578536 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.578825 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.578950 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.579053 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.579138 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:11Z","lastTransitionTime":"2025-10-14T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.682282 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.682363 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.682390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.682423 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.682445 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:11Z","lastTransitionTime":"2025-10-14T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.762263 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:11 crc kubenswrapper[4945]: E1014 15:16:11.762492 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.786303 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.786400 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.786421 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.786454 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.786475 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:11Z","lastTransitionTime":"2025-10-14T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.890218 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.890287 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.890306 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.890337 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.890359 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:11Z","lastTransitionTime":"2025-10-14T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.993557 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.993633 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.993651 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.993679 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:11 crc kubenswrapper[4945]: I1014 15:16:11.993699 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:11Z","lastTransitionTime":"2025-10-14T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.097343 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.098480 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.098585 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.098699 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.098780 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:12Z","lastTransitionTime":"2025-10-14T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.202163 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.202253 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.202270 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.202295 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.202311 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:12Z","lastTransitionTime":"2025-10-14T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.304360 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.304416 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.304438 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.304461 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.304478 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:12Z","lastTransitionTime":"2025-10-14T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.407567 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.407645 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.407669 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.407697 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.407714 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:12Z","lastTransitionTime":"2025-10-14T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.510548 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.510927 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.511071 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.511411 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.511583 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:12Z","lastTransitionTime":"2025-10-14T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.614496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.614565 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.614587 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.614612 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.614631 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:12Z","lastTransitionTime":"2025-10-14T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.717261 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.717301 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.717311 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.717329 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.717339 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:12Z","lastTransitionTime":"2025-10-14T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.762339 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.762379 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.762460 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:12 crc kubenswrapper[4945]: E1014 15:16:12.762564 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:12 crc kubenswrapper[4945]: E1014 15:16:12.762611 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:12 crc kubenswrapper[4945]: E1014 15:16:12.762709 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.819555 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.819591 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.819599 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.819612 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.819621 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:12Z","lastTransitionTime":"2025-10-14T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.922721 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.922759 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.922770 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.922787 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:12 crc kubenswrapper[4945]: I1014 15:16:12.922797 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:12Z","lastTransitionTime":"2025-10-14T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.026362 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.026419 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.026435 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.026459 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.026476 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:13Z","lastTransitionTime":"2025-10-14T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.130105 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.130155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.130175 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.130198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.130215 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:13Z","lastTransitionTime":"2025-10-14T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.233480 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.233538 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.233555 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.233589 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.233624 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:13Z","lastTransitionTime":"2025-10-14T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.337076 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.337127 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.337147 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.337170 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.337188 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:13Z","lastTransitionTime":"2025-10-14T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.440082 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.440171 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.440190 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.440216 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.440233 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:13Z","lastTransitionTime":"2025-10-14T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.543306 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.543369 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.543392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.543425 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.543446 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:13Z","lastTransitionTime":"2025-10-14T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.646514 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.646582 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.646607 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.646639 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.646661 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:13Z","lastTransitionTime":"2025-10-14T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.750499 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.750562 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.750585 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.750613 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.750637 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:13Z","lastTransitionTime":"2025-10-14T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.762604 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:13 crc kubenswrapper[4945]: E1014 15:16:13.762934 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.854490 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.854831 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.855060 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.855249 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.855392 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:13Z","lastTransitionTime":"2025-10-14T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.958782 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.959078 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.959208 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.959309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:13 crc kubenswrapper[4945]: I1014 15:16:13.959446 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:13Z","lastTransitionTime":"2025-10-14T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.062771 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.063186 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.063227 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.063288 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.063315 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:14Z","lastTransitionTime":"2025-10-14T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.168145 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.168239 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.168262 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.168292 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.168315 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:14Z","lastTransitionTime":"2025-10-14T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.271572 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.271645 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.271664 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.271692 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.271717 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:14Z","lastTransitionTime":"2025-10-14T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.375365 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.375414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.375428 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.375446 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.375458 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:14Z","lastTransitionTime":"2025-10-14T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.479018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.479091 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.479109 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.479132 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.479149 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:14Z","lastTransitionTime":"2025-10-14T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.582598 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.582931 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.583063 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.583193 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.583321 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:14Z","lastTransitionTime":"2025-10-14T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.686336 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.686388 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.686401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.686422 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.686436 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:14Z","lastTransitionTime":"2025-10-14T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.762561 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:14 crc kubenswrapper[4945]: E1014 15:16:14.762775 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.762926 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:14 crc kubenswrapper[4945]: E1014 15:16:14.763111 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.763706 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:14 crc kubenswrapper[4945]: E1014 15:16:14.764290 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.789390 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.789706 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.790018 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.790258 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.790818 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:14Z","lastTransitionTime":"2025-10-14T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.894736 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.894805 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.894868 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.894937 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.894955 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:14Z","lastTransitionTime":"2025-10-14T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.997951 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.998263 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.998369 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.998471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:14 crc kubenswrapper[4945]: I1014 15:16:14.998564 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:14Z","lastTransitionTime":"2025-10-14T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.101743 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.101833 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.101854 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.101903 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.101927 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:15Z","lastTransitionTime":"2025-10-14T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.205350 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.205437 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.205466 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.205500 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.205522 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:15Z","lastTransitionTime":"2025-10-14T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.308625 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.308701 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.308720 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.308745 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.308763 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:15Z","lastTransitionTime":"2025-10-14T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.411529 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.411581 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.411600 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.411620 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.411636 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:15Z","lastTransitionTime":"2025-10-14T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.514309 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.514373 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.514392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.514415 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.514432 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:15Z","lastTransitionTime":"2025-10-14T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.617460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.617504 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.617521 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.617539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.617550 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:15Z","lastTransitionTime":"2025-10-14T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.721685 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.721753 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.721772 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.721805 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.721826 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:15Z","lastTransitionTime":"2025-10-14T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.761939 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:15 crc kubenswrapper[4945]: E1014 15:16:15.762127 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.825317 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.825386 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.825427 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.825460 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.825484 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:15Z","lastTransitionTime":"2025-10-14T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.928549 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.928624 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.928648 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.928681 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:15 crc kubenswrapper[4945]: I1014 15:16:15.928703 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:15Z","lastTransitionTime":"2025-10-14T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.031126 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.031185 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.031199 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.031224 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.031238 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:16Z","lastTransitionTime":"2025-10-14T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.135134 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.135198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.135214 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.135233 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.135247 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:16Z","lastTransitionTime":"2025-10-14T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.238414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.238864 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.239155 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.239346 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.239531 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:16Z","lastTransitionTime":"2025-10-14T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.343171 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.343230 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.343241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.343263 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.343276 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:16Z","lastTransitionTime":"2025-10-14T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.446365 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.446393 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.446401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.446414 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.446425 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:16Z","lastTransitionTime":"2025-10-14T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.549656 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.550090 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.550320 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.550526 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.550666 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:16Z","lastTransitionTime":"2025-10-14T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.653326 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.653392 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.653410 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.653438 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.653456 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:16Z","lastTransitionTime":"2025-10-14T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.756170 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.756211 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.756221 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.756240 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.756251 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:16Z","lastTransitionTime":"2025-10-14T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.762621 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:16 crc kubenswrapper[4945]: E1014 15:16:16.762747 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.762773 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.762779 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:16 crc kubenswrapper[4945]: E1014 15:16:16.763060 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:16 crc kubenswrapper[4945]: E1014 15:16:16.763182 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.787863 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.859294 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.859345 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.859364 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.859389 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.859409 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:16Z","lastTransitionTime":"2025-10-14T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.962216 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.962274 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.962296 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.962321 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:16 crc kubenswrapper[4945]: I1014 15:16:16.962338 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:16Z","lastTransitionTime":"2025-10-14T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.065702 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.065781 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.065799 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.065827 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.065848 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:17Z","lastTransitionTime":"2025-10-14T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.169485 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.169539 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.169557 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.169580 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.169597 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:17Z","lastTransitionTime":"2025-10-14T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.271998 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.272062 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.272075 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.272093 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.272106 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:17Z","lastTransitionTime":"2025-10-14T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.373979 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.374023 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.374039 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.374055 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.374066 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:17Z","lastTransitionTime":"2025-10-14T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.478323 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.478384 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.478401 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.478429 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.478446 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:17Z","lastTransitionTime":"2025-10-14T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.582032 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.582082 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.582098 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.582122 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.582134 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:17Z","lastTransitionTime":"2025-10-14T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.685048 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.685107 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.685119 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.685143 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.685157 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:17Z","lastTransitionTime":"2025-10-14T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.762109 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:17 crc kubenswrapper[4945]: E1014 15:16:17.762307 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.787186 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.787238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.787254 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.787277 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.787293 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:17Z","lastTransitionTime":"2025-10-14T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.890534 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.890611 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.890626 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.890651 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.890668 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:17Z","lastTransitionTime":"2025-10-14T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.994500 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.994557 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.994579 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.994597 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:17 crc kubenswrapper[4945]: I1014 15:16:17.994757 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:17Z","lastTransitionTime":"2025-10-14T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.097521 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.097573 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.097587 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.097605 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.097617 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:18Z","lastTransitionTime":"2025-10-14T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.200496 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.200545 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.200557 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.200577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.200589 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:18Z","lastTransitionTime":"2025-10-14T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.303646 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.303709 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.303720 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.303738 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.303750 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:18Z","lastTransitionTime":"2025-10-14T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.406361 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.406442 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.406468 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.406500 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.406524 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:18Z","lastTransitionTime":"2025-10-14T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.508592 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.508643 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.508655 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.508672 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.508685 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:18Z","lastTransitionTime":"2025-10-14T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.611241 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.611291 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.611305 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.611324 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.611337 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:18Z","lastTransitionTime":"2025-10-14T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.713684 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.713723 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.713734 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.713749 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.713758 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:18Z","lastTransitionTime":"2025-10-14T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.762369 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.762464 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:18 crc kubenswrapper[4945]: E1014 15:16:18.762498 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.762565 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:18 crc kubenswrapper[4945]: E1014 15:16:18.762621 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:18 crc kubenswrapper[4945]: E1014 15:16:18.762743 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.816198 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.816238 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.816249 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.816265 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.816277 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:18Z","lastTransitionTime":"2025-10-14T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.919519 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.919587 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.919601 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.919629 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:18 crc kubenswrapper[4945]: I1014 15:16:18.919647 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:18Z","lastTransitionTime":"2025-10-14T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.022471 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.022524 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.022542 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.022564 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.022586 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:19Z","lastTransitionTime":"2025-10-14T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.101577 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.101748 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.101781 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.101809 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.101832 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:19Z","lastTransitionTime":"2025-10-14T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.126385 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.126455 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.126473 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.126499 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.126518 4945 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T15:16:19Z","lastTransitionTime":"2025-10-14T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.154416 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955"] Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.155114 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.157085 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.157201 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.157297 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.158117 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.200455 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.2004344160000002 podStartE2EDuration="3.200434416s" podCreationTimestamp="2025-10-14 15:16:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:19.198990966 +0000 UTC m=+109.183039344" watchObservedRunningTime="2025-10-14 15:16:19.200434416 +0000 UTC m=+109.184482794" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.272443 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.272508 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.272563 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.272592 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.272642 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.373368 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.373495 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.373806 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.373990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.374124 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.374174 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.374183 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.376299 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.385055 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.392308 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb0cb125-a5da-41be-9ffb-b1c8847e3a2b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zb955\" (UID: \"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.472823 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" Oct 14 15:16:19 crc kubenswrapper[4945]: I1014 15:16:19.762033 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:19 crc kubenswrapper[4945]: E1014 15:16:19.762182 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:20 crc kubenswrapper[4945]: I1014 15:16:20.331936 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" event={"ID":"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b","Type":"ContainerStarted","Data":"65b31736e7c05ab8e0cc627ef4b6c6d7ca6084b73a1ef7ec93fc4fc347c67747"} Oct 14 15:16:20 crc kubenswrapper[4945]: I1014 15:16:20.332000 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" event={"ID":"fb0cb125-a5da-41be-9ffb-b1c8847e3a2b","Type":"ContainerStarted","Data":"69ce70a9cd99e62f6f4e01c4cf1f75454f5ad2807d6c599f0d5dc49c8e302ac3"} Oct 14 15:16:20 crc kubenswrapper[4945]: I1014 15:16:20.353049 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zb955" podStartSLOduration=89.353015984 podStartE2EDuration="1m29.353015984s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:20.350830892 +0000 UTC m=+110.334879320" watchObservedRunningTime="2025-10-14 15:16:20.353015984 +0000 UTC m=+110.337064392" Oct 14 15:16:20 crc kubenswrapper[4945]: I1014 15:16:20.762845 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:20 crc kubenswrapper[4945]: I1014 15:16:20.762949 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:20 crc kubenswrapper[4945]: I1014 15:16:20.765000 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:20 crc kubenswrapper[4945]: E1014 15:16:20.764973 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:20 crc kubenswrapper[4945]: E1014 15:16:20.765322 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:20 crc kubenswrapper[4945]: I1014 15:16:20.765582 4945 scope.go:117] "RemoveContainer" containerID="70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134" Oct 14 15:16:20 crc kubenswrapper[4945]: E1014 15:16:20.765726 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" Oct 14 15:16:20 crc kubenswrapper[4945]: E1014 15:16:20.765974 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:21 crc kubenswrapper[4945]: I1014 15:16:21.762659 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:21 crc kubenswrapper[4945]: E1014 15:16:21.763162 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:22 crc kubenswrapper[4945]: I1014 15:16:22.763049 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:22 crc kubenswrapper[4945]: I1014 15:16:22.763737 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:22 crc kubenswrapper[4945]: I1014 15:16:22.764052 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:22 crc kubenswrapper[4945]: E1014 15:16:22.764626 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:22 crc kubenswrapper[4945]: E1014 15:16:22.764080 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:22 crc kubenswrapper[4945]: E1014 15:16:22.764850 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:23 crc kubenswrapper[4945]: I1014 15:16:23.765492 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:23 crc kubenswrapper[4945]: E1014 15:16:23.765649 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:24 crc kubenswrapper[4945]: I1014 15:16:24.762781 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:24 crc kubenswrapper[4945]: I1014 15:16:24.762820 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:24 crc kubenswrapper[4945]: I1014 15:16:24.762890 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:24 crc kubenswrapper[4945]: E1014 15:16:24.763012 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:24 crc kubenswrapper[4945]: E1014 15:16:24.763241 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:24 crc kubenswrapper[4945]: E1014 15:16:24.763402 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:25 crc kubenswrapper[4945]: I1014 15:16:25.762353 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:25 crc kubenswrapper[4945]: E1014 15:16:25.762543 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:26 crc kubenswrapper[4945]: I1014 15:16:26.350236 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b76xs_f82b9cd9-7a14-4dd9-bb0e-fde942389666/kube-multus/1.log" Oct 14 15:16:26 crc kubenswrapper[4945]: I1014 15:16:26.351067 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b76xs_f82b9cd9-7a14-4dd9-bb0e-fde942389666/kube-multus/0.log" Oct 14 15:16:26 crc kubenswrapper[4945]: I1014 15:16:26.351372 4945 generic.go:334] "Generic (PLEG): container finished" podID="f82b9cd9-7a14-4dd9-bb0e-fde942389666" containerID="17eb13e41970758fd540b4fff10d1848e4f0da616768856063641a4d1994340c" exitCode=1 Oct 14 15:16:26 crc kubenswrapper[4945]: I1014 15:16:26.351434 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b76xs" event={"ID":"f82b9cd9-7a14-4dd9-bb0e-fde942389666","Type":"ContainerDied","Data":"17eb13e41970758fd540b4fff10d1848e4f0da616768856063641a4d1994340c"} Oct 14 15:16:26 crc kubenswrapper[4945]: I1014 15:16:26.351687 4945 scope.go:117] "RemoveContainer" containerID="42024c65179c0f5377542391e6a45b1dd800e8ee91f760cbe5c0cbe55b3bd289" Oct 14 15:16:26 crc kubenswrapper[4945]: I1014 15:16:26.352177 4945 scope.go:117] "RemoveContainer" containerID="17eb13e41970758fd540b4fff10d1848e4f0da616768856063641a4d1994340c" Oct 14 15:16:26 crc kubenswrapper[4945]: E1014 15:16:26.352363 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-b76xs_openshift-multus(f82b9cd9-7a14-4dd9-bb0e-fde942389666)\"" pod="openshift-multus/multus-b76xs" podUID="f82b9cd9-7a14-4dd9-bb0e-fde942389666" Oct 14 15:16:26 crc kubenswrapper[4945]: I1014 15:16:26.762684 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:26 crc kubenswrapper[4945]: I1014 15:16:26.762755 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:26 crc kubenswrapper[4945]: E1014 15:16:26.762787 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:26 crc kubenswrapper[4945]: E1014 15:16:26.762936 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:26 crc kubenswrapper[4945]: I1014 15:16:26.762975 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:26 crc kubenswrapper[4945]: E1014 15:16:26.763044 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:27 crc kubenswrapper[4945]: I1014 15:16:27.357684 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b76xs_f82b9cd9-7a14-4dd9-bb0e-fde942389666/kube-multus/1.log" Oct 14 15:16:27 crc kubenswrapper[4945]: I1014 15:16:27.762058 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:27 crc kubenswrapper[4945]: E1014 15:16:27.762255 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:28 crc kubenswrapper[4945]: I1014 15:16:28.762388 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:28 crc kubenswrapper[4945]: E1014 15:16:28.762553 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:28 crc kubenswrapper[4945]: I1014 15:16:28.762820 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:28 crc kubenswrapper[4945]: E1014 15:16:28.762919 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:28 crc kubenswrapper[4945]: I1014 15:16:28.763183 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:28 crc kubenswrapper[4945]: E1014 15:16:28.763329 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:29 crc kubenswrapper[4945]: I1014 15:16:29.762052 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:29 crc kubenswrapper[4945]: E1014 15:16:29.762257 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:30 crc kubenswrapper[4945]: I1014 15:16:30.762565 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:30 crc kubenswrapper[4945]: I1014 15:16:30.762618 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:30 crc kubenswrapper[4945]: E1014 15:16:30.764989 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:30 crc kubenswrapper[4945]: I1014 15:16:30.765018 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:30 crc kubenswrapper[4945]: E1014 15:16:30.765664 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:30 crc kubenswrapper[4945]: E1014 15:16:30.765983 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:30 crc kubenswrapper[4945]: E1014 15:16:30.771172 4945 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 14 15:16:30 crc kubenswrapper[4945]: E1014 15:16:30.853843 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 15:16:31 crc kubenswrapper[4945]: I1014 15:16:31.762724 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:31 crc kubenswrapper[4945]: E1014 15:16:31.762983 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:32 crc kubenswrapper[4945]: I1014 15:16:32.762089 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:32 crc kubenswrapper[4945]: I1014 15:16:32.762216 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:32 crc kubenswrapper[4945]: E1014 15:16:32.762324 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:32 crc kubenswrapper[4945]: I1014 15:16:32.762351 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:32 crc kubenswrapper[4945]: E1014 15:16:32.762602 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:32 crc kubenswrapper[4945]: E1014 15:16:32.763394 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:32 crc kubenswrapper[4945]: I1014 15:16:32.763773 4945 scope.go:117] "RemoveContainer" containerID="70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134" Oct 14 15:16:32 crc kubenswrapper[4945]: E1014 15:16:32.764207 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7pqtt_openshift-ovn-kubernetes(6de8ddf6-38db-4394-8660-7e9659bf2bd2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" Oct 14 15:16:33 crc kubenswrapper[4945]: I1014 15:16:33.762457 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:33 crc kubenswrapper[4945]: E1014 15:16:33.762644 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:34 crc kubenswrapper[4945]: I1014 15:16:34.762645 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:34 crc kubenswrapper[4945]: I1014 15:16:34.762643 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:34 crc kubenswrapper[4945]: E1014 15:16:34.762972 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:34 crc kubenswrapper[4945]: I1014 15:16:34.763067 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:34 crc kubenswrapper[4945]: E1014 15:16:34.763343 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:34 crc kubenswrapper[4945]: E1014 15:16:34.763541 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:35 crc kubenswrapper[4945]: I1014 15:16:35.762305 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:35 crc kubenswrapper[4945]: E1014 15:16:35.763214 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:35 crc kubenswrapper[4945]: E1014 15:16:35.855364 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 15:16:36 crc kubenswrapper[4945]: I1014 15:16:36.762961 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:36 crc kubenswrapper[4945]: I1014 15:16:36.762980 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:36 crc kubenswrapper[4945]: E1014 15:16:36.763090 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:36 crc kubenswrapper[4945]: E1014 15:16:36.763195 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:36 crc kubenswrapper[4945]: I1014 15:16:36.762980 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:36 crc kubenswrapper[4945]: E1014 15:16:36.763267 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:37 crc kubenswrapper[4945]: I1014 15:16:37.762732 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:37 crc kubenswrapper[4945]: E1014 15:16:37.763255 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:38 crc kubenswrapper[4945]: I1014 15:16:38.762822 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:38 crc kubenswrapper[4945]: E1014 15:16:38.763024 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:38 crc kubenswrapper[4945]: I1014 15:16:38.763092 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:38 crc kubenswrapper[4945]: I1014 15:16:38.763188 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:38 crc kubenswrapper[4945]: E1014 15:16:38.763243 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:38 crc kubenswrapper[4945]: E1014 15:16:38.763685 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:39 crc kubenswrapper[4945]: I1014 15:16:39.762286 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:39 crc kubenswrapper[4945]: E1014 15:16:39.762432 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:40 crc kubenswrapper[4945]: I1014 15:16:40.762440 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:40 crc kubenswrapper[4945]: I1014 15:16:40.762459 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:40 crc kubenswrapper[4945]: I1014 15:16:40.762643 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:40 crc kubenswrapper[4945]: E1014 15:16:40.764558 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:40 crc kubenswrapper[4945]: E1014 15:16:40.764766 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:40 crc kubenswrapper[4945]: E1014 15:16:40.764836 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:40 crc kubenswrapper[4945]: E1014 15:16:40.856111 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 15:16:41 crc kubenswrapper[4945]: I1014 15:16:41.762838 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:41 crc kubenswrapper[4945]: E1014 15:16:41.763440 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:41 crc kubenswrapper[4945]: I1014 15:16:41.763469 4945 scope.go:117] "RemoveContainer" containerID="17eb13e41970758fd540b4fff10d1848e4f0da616768856063641a4d1994340c" Oct 14 15:16:42 crc kubenswrapper[4945]: I1014 15:16:42.405574 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b76xs_f82b9cd9-7a14-4dd9-bb0e-fde942389666/kube-multus/1.log" Oct 14 15:16:42 crc kubenswrapper[4945]: I1014 15:16:42.405628 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b76xs" event={"ID":"f82b9cd9-7a14-4dd9-bb0e-fde942389666","Type":"ContainerStarted","Data":"6daedef21246bc41a36b48797081a15d5a44d4a71e4e65c78f31574913dcdb57"} Oct 14 15:16:42 crc kubenswrapper[4945]: I1014 15:16:42.762789 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:42 crc kubenswrapper[4945]: I1014 15:16:42.762823 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:42 crc kubenswrapper[4945]: I1014 15:16:42.762790 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:42 crc kubenswrapper[4945]: E1014 15:16:42.762921 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:42 crc kubenswrapper[4945]: E1014 15:16:42.763085 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:42 crc kubenswrapper[4945]: E1014 15:16:42.763129 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:43 crc kubenswrapper[4945]: I1014 15:16:43.762570 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:43 crc kubenswrapper[4945]: E1014 15:16:43.762922 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:44 crc kubenswrapper[4945]: I1014 15:16:44.763121 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:44 crc kubenswrapper[4945]: I1014 15:16:44.763307 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:44 crc kubenswrapper[4945]: E1014 15:16:44.763312 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:44 crc kubenswrapper[4945]: I1014 15:16:44.763440 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:44 crc kubenswrapper[4945]: E1014 15:16:44.763582 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:44 crc kubenswrapper[4945]: E1014 15:16:44.763670 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:45 crc kubenswrapper[4945]: I1014 15:16:45.762493 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:45 crc kubenswrapper[4945]: E1014 15:16:45.762696 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:45 crc kubenswrapper[4945]: I1014 15:16:45.763517 4945 scope.go:117] "RemoveContainer" containerID="70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134" Oct 14 15:16:45 crc kubenswrapper[4945]: E1014 15:16:45.857771 4945 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 15:16:46 crc kubenswrapper[4945]: I1014 15:16:46.420964 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/3.log" Oct 14 15:16:46 crc kubenswrapper[4945]: I1014 15:16:46.422975 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerStarted","Data":"b7e9738cf3f80c415296bbc75a8735490f3d6112c9285ca52958a2fc67bda65e"} Oct 14 15:16:46 crc kubenswrapper[4945]: I1014 15:16:46.423387 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:16:46 crc kubenswrapper[4945]: I1014 15:16:46.455388 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podStartSLOduration=115.455372486 podStartE2EDuration="1m55.455372486s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:16:46.453619976 +0000 UTC m=+136.437668344" watchObservedRunningTime="2025-10-14 15:16:46.455372486 +0000 UTC m=+136.439420854" Oct 14 15:16:46 crc kubenswrapper[4945]: I1014 15:16:46.762724 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:46 crc kubenswrapper[4945]: I1014 15:16:46.762761 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:46 crc kubenswrapper[4945]: I1014 15:16:46.762790 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:46 crc kubenswrapper[4945]: E1014 15:16:46.762946 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:46 crc kubenswrapper[4945]: E1014 15:16:46.763040 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:46 crc kubenswrapper[4945]: E1014 15:16:46.763119 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:46 crc kubenswrapper[4945]: I1014 15:16:46.808744 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-ftg5h"] Oct 14 15:16:46 crc kubenswrapper[4945]: I1014 15:16:46.808956 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:46 crc kubenswrapper[4945]: E1014 15:16:46.809069 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:48 crc kubenswrapper[4945]: I1014 15:16:48.762931 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:48 crc kubenswrapper[4945]: I1014 15:16:48.762979 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:48 crc kubenswrapper[4945]: E1014 15:16:48.763554 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:48 crc kubenswrapper[4945]: I1014 15:16:48.763205 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:48 crc kubenswrapper[4945]: I1014 15:16:48.763043 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:48 crc kubenswrapper[4945]: E1014 15:16:48.763663 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:48 crc kubenswrapper[4945]: E1014 15:16:48.763762 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:48 crc kubenswrapper[4945]: E1014 15:16:48.763977 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:50 crc kubenswrapper[4945]: I1014 15:16:50.762589 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:50 crc kubenswrapper[4945]: I1014 15:16:50.762719 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:50 crc kubenswrapper[4945]: I1014 15:16:50.762742 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:50 crc kubenswrapper[4945]: I1014 15:16:50.762806 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:50 crc kubenswrapper[4945]: E1014 15:16:50.764804 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 15:16:50 crc kubenswrapper[4945]: E1014 15:16:50.765054 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 15:16:50 crc kubenswrapper[4945]: E1014 15:16:50.765213 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 15:16:50 crc kubenswrapper[4945]: E1014 15:16:50.765400 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ftg5h" podUID="be17bc84-909e-4150-8ee7-9a7d23173ba0" Oct 14 15:16:52 crc kubenswrapper[4945]: I1014 15:16:52.761951 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:52 crc kubenswrapper[4945]: I1014 15:16:52.762036 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:52 crc kubenswrapper[4945]: I1014 15:16:52.762217 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:16:52 crc kubenswrapper[4945]: I1014 15:16:52.762270 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:52 crc kubenswrapper[4945]: I1014 15:16:52.766950 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 14 15:16:52 crc kubenswrapper[4945]: I1014 15:16:52.766953 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 14 15:16:52 crc kubenswrapper[4945]: I1014 15:16:52.767833 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 14 15:16:52 crc kubenswrapper[4945]: I1014 15:16:52.767849 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 14 15:16:52 crc kubenswrapper[4945]: I1014 15:16:52.769984 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 14 15:16:52 crc kubenswrapper[4945]: I1014 15:16:52.770191 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 14 15:16:58 crc kubenswrapper[4945]: I1014 15:16:58.620474 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:16:58 crc kubenswrapper[4945]: E1014 15:16:58.620726 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:19:00.620689443 +0000 UTC m=+270.604737841 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:16:58 crc kubenswrapper[4945]: I1014 15:16:58.621836 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:58 crc kubenswrapper[4945]: I1014 15:16:58.621951 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:58 crc kubenswrapper[4945]: I1014 15:16:58.626546 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:58 crc kubenswrapper[4945]: I1014 15:16:58.630956 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:58 crc kubenswrapper[4945]: I1014 15:16:58.722987 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:58 crc kubenswrapper[4945]: I1014 15:16:58.723068 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:58 crc kubenswrapper[4945]: I1014 15:16:58.729504 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:58 crc kubenswrapper[4945]: I1014 15:16:58.729624 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:58 crc kubenswrapper[4945]: I1014 15:16:58.793611 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 15:16:58 crc kubenswrapper[4945]: I1014 15:16:58.805425 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 15:16:58 crc kubenswrapper[4945]: I1014 15:16:58.820746 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:59 crc kubenswrapper[4945]: W1014 15:16:59.005550 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-b7b5959202e4046e225869c1bae90332d4e01a87c1d12415a48ca1c91324ccdc WatchSource:0}: Error finding container b7b5959202e4046e225869c1bae90332d4e01a87c1d12415a48ca1c91324ccdc: Status 404 returned error can't find the container with id b7b5959202e4046e225869c1bae90332d4e01a87c1d12415a48ca1c91324ccdc Oct 14 15:16:59 crc kubenswrapper[4945]: W1014 15:16:59.020309 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-c3312088cb5b874bda9a46b91df3c5227cb2edeb2c1f5e2cf080bbd56695f2b4 WatchSource:0}: Error finding container c3312088cb5b874bda9a46b91df3c5227cb2edeb2c1f5e2cf080bbd56695f2b4: Status 404 returned error can't find the container with id c3312088cb5b874bda9a46b91df3c5227cb2edeb2c1f5e2cf080bbd56695f2b4 Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.474751 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e6167ccd1e141df08986e2078b8d8bf93376a68bdea3a77491df92a63475522c"} Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.475200 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6bb9bdfce5378b754b9eaebc469baf3aedd96e7f06e067617e7c4a4b92579323"} Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.475410 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.476842 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"0ba2879d85d50d9726f17d7994a13568a11ea47208acfe70076afad31b564522"} Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.476905 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c3312088cb5b874bda9a46b91df3c5227cb2edeb2c1f5e2cf080bbd56695f2b4"} Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.479538 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5d747e5f8b158e1b96ea92c9e916cb9f5c45414b3b2866fe409cb988a7dac228"} Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.479779 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b7b5959202e4046e225869c1bae90332d4e01a87c1d12415a48ca1c91324ccdc"} Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.882709 4945 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.937883 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.938388 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.938612 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-l72k8"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.939207 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-l72k8" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.940648 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fkbmp"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.945968 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.946056 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.947379 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.947416 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.949898 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.949785 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.950124 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.951860 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.953344 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.953749 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.954935 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.963862 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.965175 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.965473 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.965655 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.965766 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.965940 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.966041 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.966200 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.966274 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-frj5t"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.966688 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.967252 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.967349 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-glv7n"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.967429 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.967577 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.967673 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.970126 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.971025 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.971542 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9zzcq"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.974145 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.977741 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.978000 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.978288 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.978340 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.978569 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.978724 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.978305 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.980297 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-ccxsw"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.980903 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.980924 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bvplf"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.981066 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.981074 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.981146 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.981705 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.982678 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.986544 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.986920 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hvp57"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.987448 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.988620 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8f2dm"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.989527 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.993285 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.993434 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.993549 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.993697 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.993740 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.993867 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.993945 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.994031 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.994331 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.994463 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.994639 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.994849 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.994956 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995064 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995128 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995286 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995332 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995537 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995623 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995290 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995681 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995739 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995807 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995893 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995943 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.995625 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.996047 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.996329 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.996438 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.996560 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.996600 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-6cv87"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.996664 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.996759 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.996987 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6cv87" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.997442 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.997585 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.997693 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.997821 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.997967 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.998135 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rm8t5"] Oct 14 15:16:59 crc kubenswrapper[4945]: I1014 15:16:59.998412 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.000085 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.000675 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.001634 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.002251 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.004146 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.010760 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.013002 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.013424 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.013649 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.013771 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.013889 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4bf5h"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.014614 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.014666 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.014710 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.020066 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.014844 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.014911 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.020961 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.023339 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.026637 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.030530 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.030530 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.035271 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.035600 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.035783 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.038995 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.041532 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.041777 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.041829 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7ff096c7-1a05-493a-9e19-8bec41303e6c-client-ca\") pod \"route-controller-manager-6576b87f9c-hxtjg\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.041859 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c3400fe-3438-4116-b141-3281b0fa2d75-config-volume\") pod \"dns-default-l72k8\" (UID: \"2c3400fe-3438-4116-b141-3281b0fa2d75\") " pod="openshift-dns/dns-default-l72k8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.041902 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lzp8\" (UniqueName: \"kubernetes.io/projected/2a40af83-59c4-4c08-a839-4274b3ef62ba-kube-api-access-6lzp8\") pod \"machine-approver-56656f9798-pkb5l\" (UID: \"2a40af83-59c4-4c08-a839-4274b3ef62ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.041920 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042066 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.041923 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff096c7-1a05-493a-9e19-8bec41303e6c-serving-cert\") pod \"route-controller-manager-6576b87f9c-hxtjg\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042230 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/932ab187-513c-4d0f-9c6f-f986ac444f78-images\") pod \"machine-api-operator-5694c8668f-fkbmp\" (UID: \"932ab187-513c-4d0f-9c6f-f986ac444f78\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042285 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/932ab187-513c-4d0f-9c6f-f986ac444f78-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fkbmp\" (UID: \"932ab187-513c-4d0f-9c6f-f986ac444f78\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042345 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a40af83-59c4-4c08-a839-4274b3ef62ba-config\") pod \"machine-approver-56656f9798-pkb5l\" (UID: \"2a40af83-59c4-4c08-a839-4274b3ef62ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042367 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932ab187-513c-4d0f-9c6f-f986ac444f78-config\") pod \"machine-api-operator-5694c8668f-fkbmp\" (UID: \"932ab187-513c-4d0f-9c6f-f986ac444f78\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042385 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ff096c7-1a05-493a-9e19-8bec41303e6c-config\") pod \"route-controller-manager-6576b87f9c-hxtjg\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042416 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7fv7\" (UniqueName: \"kubernetes.io/projected/638d7074-0fbb-4086-9c1d-37036bee9034-kube-api-access-h7fv7\") pod \"openshift-apiserver-operator-796bbdcf4f-78cw8\" (UID: \"638d7074-0fbb-4086-9c1d-37036bee9034\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042459 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7jdq\" (UniqueName: \"kubernetes.io/projected/7ff096c7-1a05-493a-9e19-8bec41303e6c-kube-api-access-s7jdq\") pod \"route-controller-manager-6576b87f9c-hxtjg\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042478 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/638d7074-0fbb-4086-9c1d-37036bee9034-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-78cw8\" (UID: \"638d7074-0fbb-4086-9c1d-37036bee9034\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042495 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvrfl\" (UniqueName: \"kubernetes.io/projected/2c3400fe-3438-4116-b141-3281b0fa2d75-kube-api-access-kvrfl\") pod \"dns-default-l72k8\" (UID: \"2c3400fe-3438-4116-b141-3281b0fa2d75\") " pod="openshift-dns/dns-default-l72k8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042512 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2a40af83-59c4-4c08-a839-4274b3ef62ba-machine-approver-tls\") pod \"machine-approver-56656f9798-pkb5l\" (UID: \"2a40af83-59c4-4c08-a839-4274b3ef62ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042530 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/638d7074-0fbb-4086-9c1d-37036bee9034-config\") pod \"openshift-apiserver-operator-796bbdcf4f-78cw8\" (UID: \"638d7074-0fbb-4086-9c1d-37036bee9034\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042555 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjphb\" (UniqueName: \"kubernetes.io/projected/932ab187-513c-4d0f-9c6f-f986ac444f78-kube-api-access-cjphb\") pod \"machine-api-operator-5694c8668f-fkbmp\" (UID: \"932ab187-513c-4d0f-9c6f-f986ac444f78\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042582 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2a40af83-59c4-4c08-a839-4274b3ef62ba-auth-proxy-config\") pod \"machine-approver-56656f9798-pkb5l\" (UID: \"2a40af83-59c4-4c08-a839-4274b3ef62ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042602 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2c3400fe-3438-4116-b141-3281b0fa2d75-metrics-tls\") pod \"dns-default-l72k8\" (UID: \"2c3400fe-3438-4116-b141-3281b0fa2d75\") " pod="openshift-dns/dns-default-l72k8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042724 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042866 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042955 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.043059 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.043076 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.042893 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.057184 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.057322 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.058643 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.059139 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.067331 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.067620 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.067798 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.068150 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.068895 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.069735 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.070065 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.070591 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.070911 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fkbmp"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.071117 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.071788 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.071969 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.073680 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.074080 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.074092 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.074312 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.075129 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.077818 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.078398 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.079327 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.080090 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.082766 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.082993 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-45pg9"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.084170 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.084477 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-45pg9" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.084505 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-l72k8"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.084625 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.085774 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-6j5pp"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.086385 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.086927 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.087107 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.087417 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.087826 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.101354 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-c25w8"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.101748 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.102805 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.104478 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.106562 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.106556 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.106717 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.108862 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.112690 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-frj5t"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.112753 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.134399 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.135991 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2kwtj"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.136617 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9chhc"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.137008 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.137127 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.137261 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.137627 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.139269 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.139937 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.140108 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-s5dk6"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146425 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/638d7074-0fbb-4086-9c1d-37036bee9034-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-78cw8\" (UID: \"638d7074-0fbb-4086-9c1d-37036bee9034\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146460 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvrfl\" (UniqueName: \"kubernetes.io/projected/2c3400fe-3438-4116-b141-3281b0fa2d75-kube-api-access-kvrfl\") pod \"dns-default-l72k8\" (UID: \"2c3400fe-3438-4116-b141-3281b0fa2d75\") " pod="openshift-dns/dns-default-l72k8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146500 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7jdq\" (UniqueName: \"kubernetes.io/projected/7ff096c7-1a05-493a-9e19-8bec41303e6c-kube-api-access-s7jdq\") pod \"route-controller-manager-6576b87f9c-hxtjg\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146525 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d55256a5-30da-4fd3-beef-2d3c785ffaa4-serving-cert\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146541 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2a40af83-59c4-4c08-a839-4274b3ef62ba-machine-approver-tls\") pod \"machine-approver-56656f9798-pkb5l\" (UID: \"2a40af83-59c4-4c08-a839-4274b3ef62ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146558 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146578 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/638d7074-0fbb-4086-9c1d-37036bee9034-config\") pod \"openshift-apiserver-operator-796bbdcf4f-78cw8\" (UID: \"638d7074-0fbb-4086-9c1d-37036bee9034\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146594 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7btq\" (UniqueName: \"kubernetes.io/projected/50be8f46-65da-4733-bb6d-b4dc67c03b60-kube-api-access-w7btq\") pod \"openshift-config-operator-7777fb866f-bvplf\" (UID: \"50be8f46-65da-4733-bb6d-b4dc67c03b60\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146613 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s8m6\" (UniqueName: \"kubernetes.io/projected/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-kube-api-access-2s8m6\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146630 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b1593874-4de8-4e33-aadb-0dbdf30e7262-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5cj52\" (UID: \"b1593874-4de8-4e33-aadb-0dbdf30e7262\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146650 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjphb\" (UniqueName: \"kubernetes.io/projected/932ab187-513c-4d0f-9c6f-f986ac444f78-kube-api-access-cjphb\") pod \"machine-api-operator-5694c8668f-fkbmp\" (UID: \"932ab187-513c-4d0f-9c6f-f986ac444f78\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146666 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146683 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50be8f46-65da-4733-bb6d-b4dc67c03b60-serving-cert\") pod \"openshift-config-operator-7777fb866f-bvplf\" (UID: \"50be8f46-65da-4733-bb6d-b4dc67c03b60\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146700 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-service-ca-bundle\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146713 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-audit-policies\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146730 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l29x4\" (UniqueName: \"kubernetes.io/projected/b1593874-4de8-4e33-aadb-0dbdf30e7262-kube-api-access-l29x4\") pod \"cluster-image-registry-operator-dc59b4c8b-5cj52\" (UID: \"b1593874-4de8-4e33-aadb-0dbdf30e7262\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146749 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/388c7278-fae0-4de0-b646-49e46f88c517-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8q5kj\" (UID: \"388c7278-fae0-4de0-b646-49e46f88c517\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146764 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-config\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146779 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2a40af83-59c4-4c08-a839-4274b3ef62ba-auth-proxy-config\") pod \"machine-approver-56656f9798-pkb5l\" (UID: \"2a40af83-59c4-4c08-a839-4274b3ef62ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146795 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146811 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146828 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8fch\" (UniqueName: \"kubernetes.io/projected/d55256a5-30da-4fd3-beef-2d3c785ffaa4-kube-api-access-n8fch\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146843 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2c3400fe-3438-4116-b141-3281b0fa2d75-metrics-tls\") pod \"dns-default-l72k8\" (UID: \"2c3400fe-3438-4116-b141-3281b0fa2d75\") " pod="openshift-dns/dns-default-l72k8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146857 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146905 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d55256a5-30da-4fd3-beef-2d3c785ffaa4-etcd-ca\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146921 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58633adf-8733-4b72-a9da-78e278ebe5fd-serving-cert\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146937 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146954 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7ff096c7-1a05-493a-9e19-8bec41303e6c-client-ca\") pod \"route-controller-manager-6576b87f9c-hxtjg\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146969 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c3400fe-3438-4116-b141-3281b0fa2d75-config-volume\") pod \"dns-default-l72k8\" (UID: \"2c3400fe-3438-4116-b141-3281b0fa2d75\") " pod="openshift-dns/dns-default-l72k8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.146985 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/82aa3bda-5c5a-4d70-826e-e0a5d358e3f2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qnvrc\" (UID: \"82aa3bda-5c5a-4d70-826e-e0a5d358e3f2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147005 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/388c7278-fae0-4de0-b646-49e46f88c517-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8q5kj\" (UID: \"388c7278-fae0-4de0-b646-49e46f88c517\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147024 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147048 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lzp8\" (UniqueName: \"kubernetes.io/projected/2a40af83-59c4-4c08-a839-4274b3ef62ba-kube-api-access-6lzp8\") pod \"machine-approver-56656f9798-pkb5l\" (UID: \"2a40af83-59c4-4c08-a839-4274b3ef62ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147067 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff096c7-1a05-493a-9e19-8bec41303e6c-serving-cert\") pod \"route-controller-manager-6576b87f9c-hxtjg\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147086 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d55256a5-30da-4fd3-beef-2d3c785ffaa4-config\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147106 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/932ab187-513c-4d0f-9c6f-f986ac444f78-images\") pod \"machine-api-operator-5694c8668f-fkbmp\" (UID: \"932ab187-513c-4d0f-9c6f-f986ac444f78\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147127 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pztn\" (UniqueName: \"kubernetes.io/projected/d094c180-e6dd-4064-aea1-4c04b43e4963-kube-api-access-2pztn\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147150 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ce14cdf-3467-4035-bb1b-23a621ed6a93-serving-cert\") pod \"console-operator-58897d9998-hvp57\" (UID: \"0ce14cdf-3467-4035-bb1b-23a621ed6a93\") " pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147166 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/932ab187-513c-4d0f-9c6f-f986ac444f78-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fkbmp\" (UID: \"932ab187-513c-4d0f-9c6f-f986ac444f78\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147184 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-serving-cert\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147208 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147230 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147246 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b1593874-4de8-4e33-aadb-0dbdf30e7262-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5cj52\" (UID: \"b1593874-4de8-4e33-aadb-0dbdf30e7262\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147261 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/50be8f46-65da-4733-bb6d-b4dc67c03b60-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bvplf\" (UID: \"50be8f46-65da-4733-bb6d-b4dc67c03b60\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147279 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2svdj\" (UniqueName: \"kubernetes.io/projected/58804eb4-1016-427d-bcf6-fcfa9e8864df-kube-api-access-2svdj\") pod \"downloads-7954f5f757-6cv87\" (UID: \"58804eb4-1016-427d-bcf6-fcfa9e8864df\") " pod="openshift-console/downloads-7954f5f757-6cv87" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147300 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-config\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147317 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147331 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d094c180-e6dd-4064-aea1-4c04b43e4963-audit-dir\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147346 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0ce14cdf-3467-4035-bb1b-23a621ed6a93-trusted-ca\") pod \"console-operator-58897d9998-hvp57\" (UID: \"0ce14cdf-3467-4035-bb1b-23a621ed6a93\") " pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147371 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b1593874-4de8-4e33-aadb-0dbdf30e7262-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5cj52\" (UID: \"b1593874-4de8-4e33-aadb-0dbdf30e7262\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147387 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a40af83-59c4-4c08-a839-4274b3ef62ba-config\") pod \"machine-approver-56656f9798-pkb5l\" (UID: \"2a40af83-59c4-4c08-a839-4274b3ef62ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147402 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147419 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d55256a5-30da-4fd3-beef-2d3c785ffaa4-etcd-service-ca\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147434 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932ab187-513c-4d0f-9c6f-f986ac444f78-config\") pod \"machine-api-operator-5694c8668f-fkbmp\" (UID: \"932ab187-513c-4d0f-9c6f-f986ac444f78\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147448 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ff096c7-1a05-493a-9e19-8bec41303e6c-config\") pod \"route-controller-manager-6576b87f9c-hxtjg\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147463 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d55256a5-30da-4fd3-beef-2d3c785ffaa4-etcd-client\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147491 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147508 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlkkv\" (UniqueName: \"kubernetes.io/projected/388c7278-fae0-4de0-b646-49e46f88c517-kube-api-access-vlkkv\") pod \"openshift-controller-manager-operator-756b6f6bc6-8q5kj\" (UID: \"388c7278-fae0-4de0-b646-49e46f88c517\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147526 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqs6h\" (UniqueName: \"kubernetes.io/projected/82aa3bda-5c5a-4d70-826e-e0a5d358e3f2-kube-api-access-tqs6h\") pod \"cluster-samples-operator-665b6dd947-qnvrc\" (UID: \"82aa3bda-5c5a-4d70-826e-e0a5d358e3f2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147542 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq8ht\" (UniqueName: \"kubernetes.io/projected/58633adf-8733-4b72-a9da-78e278ebe5fd-kube-api-access-xq8ht\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147556 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7llw\" (UniqueName: \"kubernetes.io/projected/0ce14cdf-3467-4035-bb1b-23a621ed6a93-kube-api-access-p7llw\") pod \"console-operator-58897d9998-hvp57\" (UID: \"0ce14cdf-3467-4035-bb1b-23a621ed6a93\") " pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147573 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7fv7\" (UniqueName: \"kubernetes.io/projected/638d7074-0fbb-4086-9c1d-37036bee9034-kube-api-access-h7fv7\") pod \"openshift-apiserver-operator-796bbdcf4f-78cw8\" (UID: \"638d7074-0fbb-4086-9c1d-37036bee9034\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147587 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-client-ca\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147603 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.147620 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ce14cdf-3467-4035-bb1b-23a621ed6a93-config\") pod \"console-operator-58897d9998-hvp57\" (UID: \"0ce14cdf-3467-4035-bb1b-23a621ed6a93\") " pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.155229 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.155315 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-h98fq"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.155992 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8f2dm"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.155995 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/932ab187-513c-4d0f-9c6f-f986ac444f78-images\") pod \"machine-api-operator-5694c8668f-fkbmp\" (UID: \"932ab187-513c-4d0f-9c6f-f986ac444f78\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.156069 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-h98fq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.156229 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-s5dk6" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.156504 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c3400fe-3438-4116-b141-3281b0fa2d75-config-volume\") pod \"dns-default-l72k8\" (UID: \"2c3400fe-3438-4116-b141-3281b0fa2d75\") " pod="openshift-dns/dns-default-l72k8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.156699 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/638d7074-0fbb-4086-9c1d-37036bee9034-config\") pod \"openshift-apiserver-operator-796bbdcf4f-78cw8\" (UID: \"638d7074-0fbb-4086-9c1d-37036bee9034\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.156758 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/638d7074-0fbb-4086-9c1d-37036bee9034-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-78cw8\" (UID: \"638d7074-0fbb-4086-9c1d-37036bee9034\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.157224 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7ff096c7-1a05-493a-9e19-8bec41303e6c-client-ca\") pod \"route-controller-manager-6576b87f9c-hxtjg\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.157339 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932ab187-513c-4d0f-9c6f-f986ac444f78-config\") pod \"machine-api-operator-5694c8668f-fkbmp\" (UID: \"932ab187-513c-4d0f-9c6f-f986ac444f78\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.157559 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2a40af83-59c4-4c08-a839-4274b3ef62ba-auth-proxy-config\") pod \"machine-approver-56656f9798-pkb5l\" (UID: \"2a40af83-59c4-4c08-a839-4274b3ef62ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.157630 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2c3400fe-3438-4116-b141-3281b0fa2d75-metrics-tls\") pod \"dns-default-l72k8\" (UID: \"2c3400fe-3438-4116-b141-3281b0fa2d75\") " pod="openshift-dns/dns-default-l72k8" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.157667 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ff096c7-1a05-493a-9e19-8bec41303e6c-config\") pod \"route-controller-manager-6576b87f9c-hxtjg\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.159018 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/932ab187-513c-4d0f-9c6f-f986ac444f78-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fkbmp\" (UID: \"932ab187-513c-4d0f-9c6f-f986ac444f78\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.162448 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff096c7-1a05-493a-9e19-8bec41303e6c-serving-cert\") pod \"route-controller-manager-6576b87f9c-hxtjg\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.162780 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a40af83-59c4-4c08-a839-4274b3ef62ba-config\") pod \"machine-approver-56656f9798-pkb5l\" (UID: \"2a40af83-59c4-4c08-a839-4274b3ef62ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.164847 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2a40af83-59c4-4c08-a839-4274b3ef62ba-machine-approver-tls\") pod \"machine-approver-56656f9798-pkb5l\" (UID: \"2a40af83-59c4-4c08-a839-4274b3ef62ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.167588 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.167733 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bvplf"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.167765 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ccxsw"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.168436 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.171938 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-8zrgd"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.172547 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.172624 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8zrgd" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.175459 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.177394 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-glv7n"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.178387 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6cv87"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.180211 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4bf5h"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.181099 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.181110 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9zzcq"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.183920 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.184807 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.185988 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.186764 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hvp57"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.188068 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-tgs6r"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.189095 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.189104 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.190379 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.191531 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-c25w8"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.192744 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.194243 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.195080 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.196375 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2kwtj"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.197451 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rm8t5"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.198603 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-45pg9"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.199823 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.200684 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.200806 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.201982 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.203301 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.204528 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-s5dk6"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.205942 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9chhc"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.206792 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-h98fq"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.207988 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-tgs6r"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.208971 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-2j7ss"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.209758 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2j7ss" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.210079 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2j7ss"] Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.220614 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.241662 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248129 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248164 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248181 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8fch\" (UniqueName: \"kubernetes.io/projected/d55256a5-30da-4fd3-beef-2d3c785ffaa4-kube-api-access-n8fch\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248198 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248214 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d55256a5-30da-4fd3-beef-2d3c785ffaa4-etcd-ca\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248228 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58633adf-8733-4b72-a9da-78e278ebe5fd-serving-cert\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248244 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248262 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/82aa3bda-5c5a-4d70-826e-e0a5d358e3f2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qnvrc\" (UID: \"82aa3bda-5c5a-4d70-826e-e0a5d358e3f2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248280 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248303 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/388c7278-fae0-4de0-b646-49e46f88c517-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8q5kj\" (UID: \"388c7278-fae0-4de0-b646-49e46f88c517\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248332 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d55256a5-30da-4fd3-beef-2d3c785ffaa4-config\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248349 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pztn\" (UniqueName: \"kubernetes.io/projected/d094c180-e6dd-4064-aea1-4c04b43e4963-kube-api-access-2pztn\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248364 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ce14cdf-3467-4035-bb1b-23a621ed6a93-serving-cert\") pod \"console-operator-58897d9998-hvp57\" (UID: \"0ce14cdf-3467-4035-bb1b-23a621ed6a93\") " pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248380 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248396 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-serving-cert\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248412 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248428 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b1593874-4de8-4e33-aadb-0dbdf30e7262-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5cj52\" (UID: \"b1593874-4de8-4e33-aadb-0dbdf30e7262\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248444 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/50be8f46-65da-4733-bb6d-b4dc67c03b60-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bvplf\" (UID: \"50be8f46-65da-4733-bb6d-b4dc67c03b60\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248461 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2svdj\" (UniqueName: \"kubernetes.io/projected/58804eb4-1016-427d-bcf6-fcfa9e8864df-kube-api-access-2svdj\") pod \"downloads-7954f5f757-6cv87\" (UID: \"58804eb4-1016-427d-bcf6-fcfa9e8864df\") " pod="openshift-console/downloads-7954f5f757-6cv87" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248478 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-config\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248493 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248509 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d094c180-e6dd-4064-aea1-4c04b43e4963-audit-dir\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248524 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0ce14cdf-3467-4035-bb1b-23a621ed6a93-trusted-ca\") pod \"console-operator-58897d9998-hvp57\" (UID: \"0ce14cdf-3467-4035-bb1b-23a621ed6a93\") " pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248546 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b1593874-4de8-4e33-aadb-0dbdf30e7262-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5cj52\" (UID: \"b1593874-4de8-4e33-aadb-0dbdf30e7262\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248563 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248578 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d55256a5-30da-4fd3-beef-2d3c785ffaa4-etcd-service-ca\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248596 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d55256a5-30da-4fd3-beef-2d3c785ffaa4-etcd-client\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248614 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248632 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlkkv\" (UniqueName: \"kubernetes.io/projected/388c7278-fae0-4de0-b646-49e46f88c517-kube-api-access-vlkkv\") pod \"openshift-controller-manager-operator-756b6f6bc6-8q5kj\" (UID: \"388c7278-fae0-4de0-b646-49e46f88c517\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248648 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqs6h\" (UniqueName: \"kubernetes.io/projected/82aa3bda-5c5a-4d70-826e-e0a5d358e3f2-kube-api-access-tqs6h\") pod \"cluster-samples-operator-665b6dd947-qnvrc\" (UID: \"82aa3bda-5c5a-4d70-826e-e0a5d358e3f2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248670 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq8ht\" (UniqueName: \"kubernetes.io/projected/58633adf-8733-4b72-a9da-78e278ebe5fd-kube-api-access-xq8ht\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248691 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7llw\" (UniqueName: \"kubernetes.io/projected/0ce14cdf-3467-4035-bb1b-23a621ed6a93-kube-api-access-p7llw\") pod \"console-operator-58897d9998-hvp57\" (UID: \"0ce14cdf-3467-4035-bb1b-23a621ed6a93\") " pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248720 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-client-ca\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248740 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.248761 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ce14cdf-3467-4035-bb1b-23a621ed6a93-config\") pod \"console-operator-58897d9998-hvp57\" (UID: \"0ce14cdf-3467-4035-bb1b-23a621ed6a93\") " pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.249042 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d55256a5-30da-4fd3-beef-2d3c785ffaa4-serving-cert\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.249185 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.249216 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7btq\" (UniqueName: \"kubernetes.io/projected/50be8f46-65da-4733-bb6d-b4dc67c03b60-kube-api-access-w7btq\") pod \"openshift-config-operator-7777fb866f-bvplf\" (UID: \"50be8f46-65da-4733-bb6d-b4dc67c03b60\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.249458 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s8m6\" (UniqueName: \"kubernetes.io/projected/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-kube-api-access-2s8m6\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.249531 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b1593874-4de8-4e33-aadb-0dbdf30e7262-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5cj52\" (UID: \"b1593874-4de8-4e33-aadb-0dbdf30e7262\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.249562 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.249602 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-audit-policies\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.249623 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50be8f46-65da-4733-bb6d-b4dc67c03b60-serving-cert\") pod \"openshift-config-operator-7777fb866f-bvplf\" (UID: \"50be8f46-65da-4733-bb6d-b4dc67c03b60\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.249645 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-service-ca-bundle\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.249672 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l29x4\" (UniqueName: \"kubernetes.io/projected/b1593874-4de8-4e33-aadb-0dbdf30e7262-kube-api-access-l29x4\") pod \"cluster-image-registry-operator-dc59b4c8b-5cj52\" (UID: \"b1593874-4de8-4e33-aadb-0dbdf30e7262\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.249698 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/388c7278-fae0-4de0-b646-49e46f88c517-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8q5kj\" (UID: \"388c7278-fae0-4de0-b646-49e46f88c517\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.249724 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-config\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.250034 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d55256a5-30da-4fd3-beef-2d3c785ffaa4-config\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.250090 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0ce14cdf-3467-4035-bb1b-23a621ed6a93-trusted-ca\") pod \"console-operator-58897d9998-hvp57\" (UID: \"0ce14cdf-3467-4035-bb1b-23a621ed6a93\") " pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.250485 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.250559 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d55256a5-30da-4fd3-beef-2d3c785ffaa4-etcd-ca\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.250687 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.250930 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-config\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.251153 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b1593874-4de8-4e33-aadb-0dbdf30e7262-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5cj52\" (UID: \"b1593874-4de8-4e33-aadb-0dbdf30e7262\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.251464 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/50be8f46-65da-4733-bb6d-b4dc67c03b60-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bvplf\" (UID: \"50be8f46-65da-4733-bb6d-b4dc67c03b60\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.251472 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d55256a5-30da-4fd3-beef-2d3c785ffaa4-etcd-service-ca\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.252115 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.252293 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.252355 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-config\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.253619 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-client-ca\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.253690 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.253765 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d55256a5-30da-4fd3-beef-2d3c785ffaa4-etcd-client\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.254146 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.254300 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ce14cdf-3467-4035-bb1b-23a621ed6a93-config\") pod \"console-operator-58897d9998-hvp57\" (UID: \"0ce14cdf-3467-4035-bb1b-23a621ed6a93\") " pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.254702 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.254836 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.254853 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b1593874-4de8-4e33-aadb-0dbdf30e7262-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5cj52\" (UID: \"b1593874-4de8-4e33-aadb-0dbdf30e7262\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.255070 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d094c180-e6dd-4064-aea1-4c04b43e4963-audit-dir\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.255503 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.255531 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.255706 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-audit-policies\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.255990 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/82aa3bda-5c5a-4d70-826e-e0a5d358e3f2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qnvrc\" (UID: \"82aa3bda-5c5a-4d70-826e-e0a5d358e3f2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.256226 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-service-ca-bundle\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.256323 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/388c7278-fae0-4de0-b646-49e46f88c517-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8q5kj\" (UID: \"388c7278-fae0-4de0-b646-49e46f88c517\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.256437 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58633adf-8733-4b72-a9da-78e278ebe5fd-serving-cert\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.258138 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.258143 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50be8f46-65da-4733-bb6d-b4dc67c03b60-serving-cert\") pod \"openshift-config-operator-7777fb866f-bvplf\" (UID: \"50be8f46-65da-4733-bb6d-b4dc67c03b60\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.258570 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ce14cdf-3467-4035-bb1b-23a621ed6a93-serving-cert\") pod \"console-operator-58897d9998-hvp57\" (UID: \"0ce14cdf-3467-4035-bb1b-23a621ed6a93\") " pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.258963 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-serving-cert\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.259274 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.260787 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.267358 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d55256a5-30da-4fd3-beef-2d3c785ffaa4-serving-cert\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.267960 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/388c7278-fae0-4de0-b646-49e46f88c517-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8q5kj\" (UID: \"388c7278-fae0-4de0-b646-49e46f88c517\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.268151 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.281257 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.301587 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.321962 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.340715 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.360630 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.381318 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.401472 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.422652 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.481112 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.501020 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.520962 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.541260 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.562189 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.581969 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.601842 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.622576 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.641401 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.661375 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.681242 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.701609 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.722064 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.742143 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.761342 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.782450 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.801668 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.830183 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.841189 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.862215 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.881757 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.902017 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.920653 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.941256 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.960540 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 14 15:17:00 crc kubenswrapper[4945]: I1014 15:17:00.980452 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.000516 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.020680 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.042545 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.061105 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.082726 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.099082 4945 request.go:700] Waited for 1.011245525s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/secrets?fieldSelector=metadata.name%3Drouter-stats-default&limit=500&resourceVersion=0 Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.101363 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.121701 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.140787 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.162744 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.181723 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.202604 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.222042 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.241045 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.261269 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.280968 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.301132 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.321513 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.341188 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.362102 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.383463 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.400813 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.420978 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.441341 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.461208 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.481158 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.500699 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.521239 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.540416 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.561758 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.591412 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.601489 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.621568 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.642080 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.660731 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.697842 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lzp8\" (UniqueName: \"kubernetes.io/projected/2a40af83-59c4-4c08-a839-4274b3ef62ba-kube-api-access-6lzp8\") pod \"machine-approver-56656f9798-pkb5l\" (UID: \"2a40af83-59c4-4c08-a839-4274b3ef62ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.714869 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvrfl\" (UniqueName: \"kubernetes.io/projected/2c3400fe-3438-4116-b141-3281b0fa2d75-kube-api-access-kvrfl\") pod \"dns-default-l72k8\" (UID: \"2c3400fe-3438-4116-b141-3281b0fa2d75\") " pod="openshift-dns/dns-default-l72k8" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.735499 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7jdq\" (UniqueName: \"kubernetes.io/projected/7ff096c7-1a05-493a-9e19-8bec41303e6c-kube-api-access-s7jdq\") pod \"route-controller-manager-6576b87f9c-hxtjg\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.740411 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.761388 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.776770 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.781384 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.792839 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-l72k8" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.802516 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.820934 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.841174 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.862141 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.881660 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7fv7\" (UniqueName: \"kubernetes.io/projected/638d7074-0fbb-4086-9c1d-37036bee9034-kube-api-access-h7fv7\") pod \"openshift-apiserver-operator-796bbdcf4f-78cw8\" (UID: \"638d7074-0fbb-4086-9c1d-37036bee9034\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" Oct 14 15:17:01 crc kubenswrapper[4945]: W1014 15:17:01.881929 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a40af83_59c4_4c08_a839_4274b3ef62ba.slice/crio-8a8c9169ee636379b77c023626c9bbf597244ccfe96d14a2a880a1becd98df96 WatchSource:0}: Error finding container 8a8c9169ee636379b77c023626c9bbf597244ccfe96d14a2a880a1becd98df96: Status 404 returned error can't find the container with id 8a8c9169ee636379b77c023626c9bbf597244ccfe96d14a2a880a1becd98df96 Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.894120 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjphb\" (UniqueName: \"kubernetes.io/projected/932ab187-513c-4d0f-9c6f-f986ac444f78-kube-api-access-cjphb\") pod \"machine-api-operator-5694c8668f-fkbmp\" (UID: \"932ab187-513c-4d0f-9c6f-f986ac444f78\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.901550 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.920523 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.942885 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.946749 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-l72k8"] Oct 14 15:17:01 crc kubenswrapper[4945]: W1014 15:17:01.953696 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c3400fe_3438_4116_b141_3281b0fa2d75.slice/crio-5098dc9c648986be0330f92bdfd6c5e76c955448e3e4f14ce85b7ee40f64752b WatchSource:0}: Error finding container 5098dc9c648986be0330f92bdfd6c5e76c955448e3e4f14ce85b7ee40f64752b: Status 404 returned error can't find the container with id 5098dc9c648986be0330f92bdfd6c5e76c955448e3e4f14ce85b7ee40f64752b Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.961329 4945 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.980599 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg"] Oct 14 15:17:01 crc kubenswrapper[4945]: I1014 15:17:01.980720 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.000967 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.021517 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.040743 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.060983 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.083374 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.099430 4945 request.go:700] Waited for 1.851075916s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-etcd-operator/serviceaccounts/etcd-operator/token Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.110722 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.121283 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8fch\" (UniqueName: \"kubernetes.io/projected/d55256a5-30da-4fd3-beef-2d3c785ffaa4-kube-api-access-n8fch\") pod \"etcd-operator-b45778765-rm8t5\" (UID: \"d55256a5-30da-4fd3-beef-2d3c785ffaa4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.125441 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.137660 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq8ht\" (UniqueName: \"kubernetes.io/projected/58633adf-8733-4b72-a9da-78e278ebe5fd-kube-api-access-xq8ht\") pod \"controller-manager-879f6c89f-glv7n\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.161763 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlkkv\" (UniqueName: \"kubernetes.io/projected/388c7278-fae0-4de0-b646-49e46f88c517-kube-api-access-vlkkv\") pod \"openshift-controller-manager-operator-756b6f6bc6-8q5kj\" (UID: \"388c7278-fae0-4de0-b646-49e46f88c517\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.182048 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b1593874-4de8-4e33-aadb-0dbdf30e7262-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5cj52\" (UID: \"b1593874-4de8-4e33-aadb-0dbdf30e7262\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.194996 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7llw\" (UniqueName: \"kubernetes.io/projected/0ce14cdf-3467-4035-bb1b-23a621ed6a93-kube-api-access-p7llw\") pod \"console-operator-58897d9998-hvp57\" (UID: \"0ce14cdf-3467-4035-bb1b-23a621ed6a93\") " pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.208048 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.214029 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7btq\" (UniqueName: \"kubernetes.io/projected/50be8f46-65da-4733-bb6d-b4dc67c03b60-kube-api-access-w7btq\") pod \"openshift-config-operator-7777fb866f-bvplf\" (UID: \"50be8f46-65da-4733-bb6d-b4dc67c03b60\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.239266 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s8m6\" (UniqueName: \"kubernetes.io/projected/d8cdf424-f208-4e85-a001-0ffc3a1b27c2-kube-api-access-2s8m6\") pod \"authentication-operator-69f744f599-frj5t\" (UID: \"d8cdf424-f208-4e85-a001-0ffc3a1b27c2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.259178 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2svdj\" (UniqueName: \"kubernetes.io/projected/58804eb4-1016-427d-bcf6-fcfa9e8864df-kube-api-access-2svdj\") pod \"downloads-7954f5f757-6cv87\" (UID: \"58804eb4-1016-427d-bcf6-fcfa9e8864df\") " pod="openshift-console/downloads-7954f5f757-6cv87" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.275658 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqs6h\" (UniqueName: \"kubernetes.io/projected/82aa3bda-5c5a-4d70-826e-e0a5d358e3f2-kube-api-access-tqs6h\") pod \"cluster-samples-operator-665b6dd947-qnvrc\" (UID: \"82aa3bda-5c5a-4d70-826e-e0a5d358e3f2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.298584 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.302443 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l29x4\" (UniqueName: \"kubernetes.io/projected/b1593874-4de8-4e33-aadb-0dbdf30e7262-kube-api-access-l29x4\") pod \"cluster-image-registry-operator-dc59b4c8b-5cj52\" (UID: \"b1593874-4de8-4e33-aadb-0dbdf30e7262\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.306166 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.317575 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pztn\" (UniqueName: \"kubernetes.io/projected/d094c180-e6dd-4064-aea1-4c04b43e4963-kube-api-access-2pztn\") pod \"oauth-openshift-558db77b4-9zzcq\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.324838 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6cv87" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.333563 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.339346 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.345604 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fkbmp"] Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.345761 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.371857 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8"] Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.374982 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81c3c0f8-ad71-4a4b-8193-95051f481416-serving-cert\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375131 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-audit\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375160 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-oauth-serving-cert\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375243 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-registry-certificates\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375318 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57dea3fa-cff1-4b2b-9193-2eef6ac70d32-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-r8hlp\" (UID: \"57dea3fa-cff1-4b2b-9193-2eef6ac70d32\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375349 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-image-import-ca\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375445 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqc7b\" (UniqueName: \"kubernetes.io/projected/81c3c0f8-ad71-4a4b-8193-95051f481416-kube-api-access-bqc7b\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375514 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-bound-sa-token\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b683cb73-6094-43f6-b9f3-9f89691405c1-audit-policies\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375608 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b683cb73-6094-43f6-b9f3-9f89691405c1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375701 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/81c3c0f8-ad71-4a4b-8193-95051f481416-audit-dir\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375829 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/81c3c0f8-ad71-4a4b-8193-95051f481416-encryption-config\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375856 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-service-ca\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375894 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-trusted-ca\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375932 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68qjd\" (UniqueName: \"kubernetes.io/projected/b683cb73-6094-43f6-b9f3-9f89691405c1-kube-api-access-68qjd\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375955 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-serving-cert\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.375985 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgjgp\" (UniqueName: \"kubernetes.io/projected/06ecdccd-2560-48ea-a7f3-87e3e75488a0-kube-api-access-mgjgp\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376018 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-config\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376042 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57dea3fa-cff1-4b2b-9193-2eef6ac70d32-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-r8hlp\" (UID: \"57dea3fa-cff1-4b2b-9193-2eef6ac70d32\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376064 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-etcd-serving-ca\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376113 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c26cc\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-kube-api-access-c26cc\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376136 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-registry-tls\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376176 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b683cb73-6094-43f6-b9f3-9f89691405c1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376201 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b683cb73-6094-43f6-b9f3-9f89691405c1-encryption-config\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376223 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b683cb73-6094-43f6-b9f3-9f89691405c1-audit-dir\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376360 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-config\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376393 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376419 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376440 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b683cb73-6094-43f6-b9f3-9f89691405c1-serving-cert\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376478 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376508 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-oauth-config\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376549 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376840 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b683cb73-6094-43f6-b9f3-9f89691405c1-etcd-client\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376903 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/81c3c0f8-ad71-4a4b-8193-95051f481416-node-pullsecrets\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: E1014 15:17:02.376939 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:02.876921216 +0000 UTC m=+152.860969674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.376972 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-trusted-ca-bundle\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.377012 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57dea3fa-cff1-4b2b-9193-2eef6ac70d32-config\") pod \"kube-controller-manager-operator-78b949d7b-r8hlp\" (UID: \"57dea3fa-cff1-4b2b-9193-2eef6ac70d32\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.377040 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/81c3c0f8-ad71-4a4b-8193-95051f481416-etcd-client\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: W1014 15:17:02.417745 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod638d7074_0fbb_4086_9c1d_37036bee9034.slice/crio-034f450c3696ff025aa14455a59ad231edb1b2ebb434f54fd608aa8b9904a373 WatchSource:0}: Error finding container 034f450c3696ff025aa14455a59ad231edb1b2ebb434f54fd608aa8b9904a373: Status 404 returned error can't find the container with id 034f450c3696ff025aa14455a59ad231edb1b2ebb434f54fd608aa8b9904a373 Oct 14 15:17:02 crc kubenswrapper[4945]: W1014 15:17:02.424693 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod932ab187_513c_4d0f_9c6f_f986ac444f78.slice/crio-ed7e47669f339bad4fcb60fc7c29f30812fac1cba6bb9a9db8f4c2f042ae38e1 WatchSource:0}: Error finding container ed7e47669f339bad4fcb60fc7c29f30812fac1cba6bb9a9db8f4c2f042ae38e1: Status 404 returned error can't find the container with id ed7e47669f339bad4fcb60fc7c29f30812fac1cba6bb9a9db8f4c2f042ae38e1 Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.441862 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-glv7n"] Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.481482 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.482395 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:02 crc kubenswrapper[4945]: E1014 15:17:02.482553 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:02.982530679 +0000 UTC m=+152.966579047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.482756 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/81c3c0f8-ad71-4a4b-8193-95051f481416-node-pullsecrets\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.482814 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa06ee0-ace4-4549-8bf4-2d02930cf2f5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nb6gw\" (UID: \"afa06ee0-ace4-4549-8bf4-2d02930cf2f5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.482863 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkcfp\" (UniqueName: \"kubernetes.io/projected/fdb8607a-0a03-4e36-bf27-080330220d71-kube-api-access-nkcfp\") pod \"olm-operator-6b444d44fb-9qkjs\" (UID: \"fdb8607a-0a03-4e36-bf27-080330220d71\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.482920 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/81c3c0f8-ad71-4a4b-8193-95051f481416-node-pullsecrets\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.482964 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fd845cc3-5d83-4f84-9757-338b2eb40662-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dmdlz\" (UID: \"fd845cc3-5d83-4f84-9757-338b2eb40662\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483277 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/81c3c0f8-ad71-4a4b-8193-95051f481416-etcd-client\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483501 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fdb8607a-0a03-4e36-bf27-080330220d71-srv-cert\") pod \"olm-operator-6b444d44fb-9qkjs\" (UID: \"fdb8607a-0a03-4e36-bf27-080330220d71\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483531 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-plugins-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483553 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a8a1bd5-206a-4154-886b-21f1b00906a6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfgxp\" (UID: \"9a8a1bd5-206a-4154-886b-21f1b00906a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483735 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-csi-data-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483756 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/39778b3c-6ffc-417e-b9ff-50526b1fd867-srv-cert\") pod \"catalog-operator-68c6474976-mk6kt\" (UID: \"39778b3c-6ffc-417e-b9ff-50526b1fd867\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483779 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv2fz\" (UniqueName: \"kubernetes.io/projected/fd845cc3-5d83-4f84-9757-338b2eb40662-kube-api-access-vv2fz\") pod \"ingress-operator-5b745b69d9-dmdlz\" (UID: \"fd845cc3-5d83-4f84-9757-338b2eb40662\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483793 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2-serving-cert\") pod \"service-ca-operator-777779d784-c25w8\" (UID: \"877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483828 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-audit\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483895 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g926\" (UniqueName: \"kubernetes.io/projected/c70614b5-a089-4e6e-8c59-06a01aaa2c54-kube-api-access-6g926\") pod \"machine-config-server-8zrgd\" (UID: \"c70614b5-a089-4e6e-8c59-06a01aaa2c54\") " pod="openshift-machine-config-operator/machine-config-server-8zrgd" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483912 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d161354c-4c64-4c8f-a368-3462e7d3c5ac-stats-auth\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483938 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-registry-certificates\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.483969 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57dea3fa-cff1-4b2b-9193-2eef6ac70d32-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-r8hlp\" (UID: \"57dea3fa-cff1-4b2b-9193-2eef6ac70d32\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484003 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6-webhook-cert\") pod \"packageserver-d55dfcdfc-jtl86\" (UID: \"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484019 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdc28\" (UniqueName: \"kubernetes.io/projected/877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2-kube-api-access-rdc28\") pod \"service-ca-operator-777779d784-c25w8\" (UID: \"877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484036 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-image-import-ca\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484052 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqc7b\" (UniqueName: \"kubernetes.io/projected/81c3c0f8-ad71-4a4b-8193-95051f481416-kube-api-access-bqc7b\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484068 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ade216cb-9ef7-4dac-b35e-61dd3547b3c3-proxy-tls\") pod \"machine-config-operator-74547568cd-m8jbn\" (UID: \"ade216cb-9ef7-4dac-b35e-61dd3547b3c3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484087 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8952m\" (UniqueName: \"kubernetes.io/projected/a98ffecf-8060-4d01-8bfe-e1f7c8f04b46-kube-api-access-8952m\") pod \"service-ca-9c57cc56f-9chhc\" (UID: \"a98ffecf-8060-4d01-8bfe-e1f7c8f04b46\") " pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484103 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab52f1c4-7d72-4d74-a16f-d49ed094b8e8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5v9kf\" (UID: \"ab52f1c4-7d72-4d74-a16f-d49ed094b8e8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484122 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-bound-sa-token\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484138 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a8a1bd5-206a-4154-886b-21f1b00906a6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfgxp\" (UID: \"9a8a1bd5-206a-4154-886b-21f1b00906a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484175 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/81c3c0f8-ad71-4a4b-8193-95051f481416-audit-dir\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484192 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cd5588a7-14c8-4a04-bccc-a22bc3edefaa-cert\") pod \"ingress-canary-2j7ss\" (UID: \"cd5588a7-14c8-4a04-bccc-a22bc3edefaa\") " pod="openshift-ingress-canary/ingress-canary-2j7ss" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484209 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krgsh\" (UniqueName: \"kubernetes.io/projected/cd5588a7-14c8-4a04-bccc-a22bc3edefaa-kube-api-access-krgsh\") pod \"ingress-canary-2j7ss\" (UID: \"cd5588a7-14c8-4a04-bccc-a22bc3edefaa\") " pod="openshift-ingress-canary/ingress-canary-2j7ss" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484233 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th2mh\" (UniqueName: \"kubernetes.io/projected/0ee69313-9ffd-4f06-aa3a-58037bd058de-kube-api-access-th2mh\") pod \"dns-operator-744455d44c-s5dk6\" (UID: \"0ee69313-9ffd-4f06-aa3a-58037bd058de\") " pod="openshift-dns-operator/dns-operator-744455d44c-s5dk6" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484248 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khq4r\" (UniqueName: \"kubernetes.io/projected/748d0e0e-3657-4ad3-8918-c0d0773e06e6-kube-api-access-khq4r\") pod \"package-server-manager-789f6589d5-8rkjt\" (UID: \"748d0e0e-3657-4ad3-8918-c0d0773e06e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484276 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-trusted-ca\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484293 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/81c3c0f8-ad71-4a4b-8193-95051f481416-encryption-config\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484308 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-service-ca\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484326 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhzng\" (UniqueName: \"kubernetes.io/projected/b47bfeba-2bc6-4c04-bd5a-a02b10ef8478-kube-api-access-fhzng\") pod \"control-plane-machine-set-operator-78cbb6b69f-jphx9\" (UID: \"b47bfeba-2bc6-4c04-bd5a-a02b10ef8478\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484345 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/748d0e0e-3657-4ad3-8918-c0d0773e06e6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8rkjt\" (UID: \"748d0e0e-3657-4ad3-8918-c0d0773e06e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484391 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68qjd\" (UniqueName: \"kubernetes.io/projected/b683cb73-6094-43f6-b9f3-9f89691405c1-kube-api-access-68qjd\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484419 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-serving-cert\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484464 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a98ffecf-8060-4d01-8bfe-e1f7c8f04b46-signing-key\") pod \"service-ca-9c57cc56f-9chhc\" (UID: \"a98ffecf-8060-4d01-8bfe-e1f7c8f04b46\") " pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484500 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-etcd-serving-ca\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484541 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57dea3fa-cff1-4b2b-9193-2eef6ac70d32-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-r8hlp\" (UID: \"57dea3fa-cff1-4b2b-9193-2eef6ac70d32\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484583 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-registry-tls\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484601 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fd845cc3-5d83-4f84-9757-338b2eb40662-metrics-tls\") pod \"ingress-operator-5b745b69d9-dmdlz\" (UID: \"fd845cc3-5d83-4f84-9757-338b2eb40662\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484625 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b683cb73-6094-43f6-b9f3-9f89691405c1-audit-dir\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484641 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ade216cb-9ef7-4dac-b35e-61dd3547b3c3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-m8jbn\" (UID: \"ade216cb-9ef7-4dac-b35e-61dd3547b3c3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484657 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c788e539-e984-4afb-ac5f-d216466ff704-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-h98fq\" (UID: \"c788e539-e984-4afb-ac5f-d216466ff704\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h98fq" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484689 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484704 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ade216cb-9ef7-4dac-b35e-61dd3547b3c3-images\") pod \"machine-config-operator-74547568cd-m8jbn\" (UID: \"ade216cb-9ef7-4dac-b35e-61dd3547b3c3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484734 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bb2dc677-b493-45e9-a3a0-fc805b303a11-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jpztt\" (UID: \"bb2dc677-b493-45e9-a3a0-fc805b303a11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484750 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2-config\") pod \"service-ca-operator-777779d784-c25w8\" (UID: \"877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484776 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-socket-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484791 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-registration-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484817 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bb2dc677-b493-45e9-a3a0-fc805b303a11-proxy-tls\") pod \"machine-config-controller-84d6567774-jpztt\" (UID: \"bb2dc677-b493-45e9-a3a0-fc805b303a11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.484833 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4md4\" (UniqueName: \"kubernetes.io/projected/d5913ac3-e4f4-4397-bc8c-3fd7ab4fee4d-kube-api-access-w4md4\") pod \"migrator-59844c95c7-45pg9\" (UID: \"d5913ac3-e4f4-4397-bc8c-3fd7ab4fee4d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-45pg9" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.485465 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-registry-certificates\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.485497 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b47bfeba-2bc6-4c04-bd5a-a02b10ef8478-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jphx9\" (UID: \"b47bfeba-2bc6-4c04-bd5a-a02b10ef8478\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.486927 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b683cb73-6094-43f6-b9f3-9f89691405c1-audit-dir\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.487026 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69vgp\" (UniqueName: \"kubernetes.io/projected/c788e539-e984-4afb-ac5f-d216466ff704-kube-api-access-69vgp\") pod \"multus-admission-controller-857f4d67dd-h98fq\" (UID: \"c788e539-e984-4afb-ac5f-d216466ff704\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h98fq" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.487290 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b683cb73-6094-43f6-b9f3-9f89691405c1-etcd-client\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.487317 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7nzl\" (UniqueName: \"kubernetes.io/projected/afa06ee0-ace4-4549-8bf4-2d02930cf2f5-kube-api-access-c7nzl\") pod \"kube-storage-version-migrator-operator-b67b599dd-nb6gw\" (UID: \"afa06ee0-ace4-4549-8bf4-2d02930cf2f5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.487347 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c70614b5-a089-4e6e-8c59-06a01aaa2c54-node-bootstrap-token\") pod \"machine-config-server-8zrgd\" (UID: \"c70614b5-a089-4e6e-8c59-06a01aaa2c54\") " pod="openshift-machine-config-operator/machine-config-server-8zrgd" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.487385 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-trusted-ca-bundle\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.487408 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qlg4\" (UniqueName: \"kubernetes.io/projected/bb2dc677-b493-45e9-a3a0-fc805b303a11-kube-api-access-4qlg4\") pod \"machine-config-controller-84d6567774-jpztt\" (UID: \"bb2dc677-b493-45e9-a3a0-fc805b303a11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.487452 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57dea3fa-cff1-4b2b-9193-2eef6ac70d32-config\") pod \"kube-controller-manager-operator-78b949d7b-r8hlp\" (UID: \"57dea3fa-cff1-4b2b-9193-2eef6ac70d32\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.487475 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ee69313-9ffd-4f06-aa3a-58037bd058de-metrics-tls\") pod \"dns-operator-744455d44c-s5dk6\" (UID: \"0ee69313-9ffd-4f06-aa3a-58037bd058de\") " pod="openshift-dns-operator/dns-operator-744455d44c-s5dk6" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.487514 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6-apiservice-cert\") pod \"packageserver-d55dfcdfc-jtl86\" (UID: \"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.487552 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81c3c0f8-ad71-4a4b-8193-95051f481416-serving-cert\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.487578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d161354c-4c64-4c8f-a368-3462e7d3c5ac-metrics-certs\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.488457 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/81c3c0f8-ad71-4a4b-8193-95051f481416-audit-dir\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.489107 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6-tmpfs\") pod \"packageserver-d55dfcdfc-jtl86\" (UID: \"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.489156 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-mountpoint-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.489181 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-config-volume\") pod \"collect-profiles-29340915-lj4bw\" (UID: \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.489224 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab52f1c4-7d72-4d74-a16f-d49ed094b8e8-config\") pod \"kube-apiserver-operator-766d6c64bb-5v9kf\" (UID: \"ab52f1c4-7d72-4d74-a16f-d49ed094b8e8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.489263 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-oauth-serving-cert\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.489321 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/39778b3c-6ffc-417e-b9ff-50526b1fd867-profile-collector-cert\") pod \"catalog-operator-68c6474976-mk6kt\" (UID: \"39778b3c-6ffc-417e-b9ff-50526b1fd867\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.489348 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/98994fb5-4f28-4381-9ae2-4c734bf79c5c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2kwtj\" (UID: \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.489390 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d161354c-4c64-4c8f-a368-3462e7d3c5ac-default-certificate\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.489414 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tr5r\" (UniqueName: \"kubernetes.io/projected/39778b3c-6ffc-417e-b9ff-50526b1fd867-kube-api-access-7tr5r\") pod \"catalog-operator-68c6474976-mk6kt\" (UID: \"39778b3c-6ffc-417e-b9ff-50526b1fd867\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.489453 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b683cb73-6094-43f6-b9f3-9f89691405c1-audit-policies\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.489475 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b683cb73-6094-43f6-b9f3-9f89691405c1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.489500 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98994fb5-4f28-4381-9ae2-4c734bf79c5c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2kwtj\" (UID: \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.490174 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-trusted-ca\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.490207 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/81c3c0f8-ad71-4a4b-8193-95051f481416-etcd-client\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.490449 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-trusted-ca-bundle\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.491028 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-service-ca\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.491278 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b683cb73-6094-43f6-b9f3-9f89691405c1-audit-policies\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.492317 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b683cb73-6094-43f6-b9f3-9f89691405c1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.492707 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b683cb73-6094-43f6-b9f3-9f89691405c1-etcd-client\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.493035 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-serving-cert\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.494405 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-oauth-serving-cert\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.494415 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.494483 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnpc4\" (UniqueName: \"kubernetes.io/projected/98994fb5-4f28-4381-9ae2-4c734bf79c5c-kube-api-access-rnpc4\") pod \"marketplace-operator-79b997595-2kwtj\" (UID: \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.494555 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm7q8\" (UniqueName: \"kubernetes.io/projected/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-kube-api-access-wm7q8\") pod \"collect-profiles-29340915-lj4bw\" (UID: \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.494575 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab52f1c4-7d72-4d74-a16f-d49ed094b8e8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5v9kf\" (UID: \"ab52f1c4-7d72-4d74-a16f-d49ed094b8e8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.494672 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9a8a1bd5-206a-4154-886b-21f1b00906a6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfgxp\" (UID: \"9a8a1bd5-206a-4154-886b-21f1b00906a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.494768 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgjgp\" (UniqueName: \"kubernetes.io/projected/06ecdccd-2560-48ea-a7f3-87e3e75488a0-kube-api-access-mgjgp\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.494797 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afa06ee0-ace4-4549-8bf4-2d02930cf2f5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nb6gw\" (UID: \"afa06ee0-ace4-4549-8bf4-2d02930cf2f5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.494823 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv85p\" (UniqueName: \"kubernetes.io/projected/ade216cb-9ef7-4dac-b35e-61dd3547b3c3-kube-api-access-fv85p\") pod \"machine-config-operator-74547568cd-m8jbn\" (UID: \"ade216cb-9ef7-4dac-b35e-61dd3547b3c3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.494847 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-config\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.494863 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-secret-volume\") pod \"collect-profiles-29340915-lj4bw\" (UID: \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.494982 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.495311 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57dea3fa-cff1-4b2b-9193-2eef6ac70d32-config\") pod \"kube-controller-manager-operator-78b949d7b-r8hlp\" (UID: \"57dea3fa-cff1-4b2b-9193-2eef6ac70d32\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.495667 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-audit\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.495794 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c26cc\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-kube-api-access-c26cc\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.495829 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx2x4\" (UniqueName: \"kubernetes.io/projected/49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6-kube-api-access-nx2x4\") pod \"packageserver-d55dfcdfc-jtl86\" (UID: \"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496066 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-config\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496133 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b683cb73-6094-43f6-b9f3-9f89691405c1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496164 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b683cb73-6094-43f6-b9f3-9f89691405c1-encryption-config\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496291 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-config\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496640 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496675 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b683cb73-6094-43f6-b9f3-9f89691405c1-serving-cert\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496695 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c70614b5-a089-4e6e-8c59-06a01aaa2c54-certs\") pod \"machine-config-server-8zrgd\" (UID: \"c70614b5-a089-4e6e-8c59-06a01aaa2c54\") " pod="openshift-machine-config-operator/machine-config-server-8zrgd" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496731 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496749 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fdb8607a-0a03-4e36-bf27-080330220d71-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9qkjs\" (UID: \"fdb8607a-0a03-4e36-bf27-080330220d71\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496860 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-oauth-config\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496893 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znqms\" (UniqueName: \"kubernetes.io/projected/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-kube-api-access-znqms\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496910 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a98ffecf-8060-4d01-8bfe-e1f7c8f04b46-signing-cabundle\") pod \"service-ca-9c57cc56f-9chhc\" (UID: \"a98ffecf-8060-4d01-8bfe-e1f7c8f04b46\") " pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496977 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.496995 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv2cx\" (UniqueName: \"kubernetes.io/projected/d161354c-4c64-4c8f-a368-3462e7d3c5ac-kube-api-access-qv2cx\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.497012 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd845cc3-5d83-4f84-9757-338b2eb40662-trusted-ca\") pod \"ingress-operator-5b745b69d9-dmdlz\" (UID: \"fd845cc3-5d83-4f84-9757-338b2eb40662\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.497109 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d161354c-4c64-4c8f-a368-3462e7d3c5ac-service-ca-bundle\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.497108 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-image-import-ca\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: E1014 15:17:02.497442 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:02.997428279 +0000 UTC m=+152.981476717 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.498548 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-config\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.498559 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-etcd-serving-ca\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.499205 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81c3c0f8-ad71-4a4b-8193-95051f481416-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.500634 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b683cb73-6094-43f6-b9f3-9f89691405c1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.503344 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57dea3fa-cff1-4b2b-9193-2eef6ac70d32-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-r8hlp\" (UID: \"57dea3fa-cff1-4b2b-9193-2eef6ac70d32\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.503372 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/81c3c0f8-ad71-4a4b-8193-95051f481416-encryption-config\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.503579 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81c3c0f8-ad71-4a4b-8193-95051f481416-serving-cert\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.505515 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-oauth-config\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.506349 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b683cb73-6094-43f6-b9f3-9f89691405c1-encryption-config\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.507055 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.508515 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-registry-tls\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.524452 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68qjd\" (UniqueName: \"kubernetes.io/projected/b683cb73-6094-43f6-b9f3-9f89691405c1-kube-api-access-68qjd\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.524780 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b683cb73-6094-43f6-b9f3-9f89691405c1-serving-cert\") pod \"apiserver-7bbb656c7d-mnqtf\" (UID: \"b683cb73-6094-43f6-b9f3-9f89691405c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.541948 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-l72k8" event={"ID":"2c3400fe-3438-4116-b141-3281b0fa2d75","Type":"ContainerStarted","Data":"39cba62e597e8e5562999061c924b1a006dea9347de2ef5506dcc368baa327eb"} Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.542000 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-l72k8" event={"ID":"2c3400fe-3438-4116-b141-3281b0fa2d75","Type":"ContainerStarted","Data":"5098dc9c648986be0330f92bdfd6c5e76c955448e3e4f14ce85b7ee40f64752b"} Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.550191 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.552643 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" event={"ID":"638d7074-0fbb-4086-9c1d-37036bee9034","Type":"ContainerStarted","Data":"034f450c3696ff025aa14455a59ad231edb1b2ebb434f54fd608aa8b9904a373"} Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.556719 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57dea3fa-cff1-4b2b-9193-2eef6ac70d32-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-r8hlp\" (UID: \"57dea3fa-cff1-4b2b-9193-2eef6ac70d32\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.560707 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" event={"ID":"58633adf-8733-4b72-a9da-78e278ebe5fd","Type":"ContainerStarted","Data":"7a5966cb1f53b1aa26beed514ab86c478a3fac8bc1e07149cde7e14883b1b58f"} Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.562420 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-bound-sa-token\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.562515 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" event={"ID":"932ab187-513c-4d0f-9c6f-f986ac444f78","Type":"ContainerStarted","Data":"ed7e47669f339bad4fcb60fc7c29f30812fac1cba6bb9a9db8f4c2f042ae38e1"} Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.579500 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bvplf"] Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.585890 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.596021 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" event={"ID":"2a40af83-59c4-4c08-a839-4274b3ef62ba","Type":"ContainerStarted","Data":"83d0f19974a371aade68b384bf5ca90ad2df995d0b503507826d630555cb2a59"} Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.596076 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" event={"ID":"2a40af83-59c4-4c08-a839-4274b3ef62ba","Type":"ContainerStarted","Data":"77f592754497cd13967cd6c0312bf8b3e2db4462acc0f8339e4183c7e245d42f"} Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.596093 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" event={"ID":"2a40af83-59c4-4c08-a839-4274b3ef62ba","Type":"ContainerStarted","Data":"8a8c9169ee636379b77c023626c9bbf597244ccfe96d14a2a880a1becd98df96"} Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.597614 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.597761 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hvp57"] Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.597795 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cd5588a7-14c8-4a04-bccc-a22bc3edefaa-cert\") pod \"ingress-canary-2j7ss\" (UID: \"cd5588a7-14c8-4a04-bccc-a22bc3edefaa\") " pod="openshift-ingress-canary/ingress-canary-2j7ss" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.597823 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krgsh\" (UniqueName: \"kubernetes.io/projected/cd5588a7-14c8-4a04-bccc-a22bc3edefaa-kube-api-access-krgsh\") pod \"ingress-canary-2j7ss\" (UID: \"cd5588a7-14c8-4a04-bccc-a22bc3edefaa\") " pod="openshift-ingress-canary/ingress-canary-2j7ss" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.597846 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th2mh\" (UniqueName: \"kubernetes.io/projected/0ee69313-9ffd-4f06-aa3a-58037bd058de-kube-api-access-th2mh\") pod \"dns-operator-744455d44c-s5dk6\" (UID: \"0ee69313-9ffd-4f06-aa3a-58037bd058de\") " pod="openshift-dns-operator/dns-operator-744455d44c-s5dk6" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.597912 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khq4r\" (UniqueName: \"kubernetes.io/projected/748d0e0e-3657-4ad3-8918-c0d0773e06e6-kube-api-access-khq4r\") pod \"package-server-manager-789f6589d5-8rkjt\" (UID: \"748d0e0e-3657-4ad3-8918-c0d0773e06e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" Oct 14 15:17:02 crc kubenswrapper[4945]: E1014 15:17:02.597951 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:03.097931768 +0000 UTC m=+153.081980136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.597979 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhzng\" (UniqueName: \"kubernetes.io/projected/b47bfeba-2bc6-4c04-bd5a-a02b10ef8478-kube-api-access-fhzng\") pod \"control-plane-machine-set-operator-78cbb6b69f-jphx9\" (UID: \"b47bfeba-2bc6-4c04-bd5a-a02b10ef8478\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598019 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/748d0e0e-3657-4ad3-8918-c0d0773e06e6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8rkjt\" (UID: \"748d0e0e-3657-4ad3-8918-c0d0773e06e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598045 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a98ffecf-8060-4d01-8bfe-e1f7c8f04b46-signing-key\") pod \"service-ca-9c57cc56f-9chhc\" (UID: \"a98ffecf-8060-4d01-8bfe-e1f7c8f04b46\") " pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598070 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fd845cc3-5d83-4f84-9757-338b2eb40662-metrics-tls\") pod \"ingress-operator-5b745b69d9-dmdlz\" (UID: \"fd845cc3-5d83-4f84-9757-338b2eb40662\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598088 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ade216cb-9ef7-4dac-b35e-61dd3547b3c3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-m8jbn\" (UID: \"ade216cb-9ef7-4dac-b35e-61dd3547b3c3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598104 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c788e539-e984-4afb-ac5f-d216466ff704-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-h98fq\" (UID: \"c788e539-e984-4afb-ac5f-d216466ff704\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h98fq" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598123 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ade216cb-9ef7-4dac-b35e-61dd3547b3c3-images\") pod \"machine-config-operator-74547568cd-m8jbn\" (UID: \"ade216cb-9ef7-4dac-b35e-61dd3547b3c3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598142 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bb2dc677-b493-45e9-a3a0-fc805b303a11-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jpztt\" (UID: \"bb2dc677-b493-45e9-a3a0-fc805b303a11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598160 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2-config\") pod \"service-ca-operator-777779d784-c25w8\" (UID: \"877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598183 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-socket-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598198 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-registration-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598221 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bb2dc677-b493-45e9-a3a0-fc805b303a11-proxy-tls\") pod \"machine-config-controller-84d6567774-jpztt\" (UID: \"bb2dc677-b493-45e9-a3a0-fc805b303a11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598237 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4md4\" (UniqueName: \"kubernetes.io/projected/d5913ac3-e4f4-4397-bc8c-3fd7ab4fee4d-kube-api-access-w4md4\") pod \"migrator-59844c95c7-45pg9\" (UID: \"d5913ac3-e4f4-4397-bc8c-3fd7ab4fee4d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-45pg9" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598256 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b47bfeba-2bc6-4c04-bd5a-a02b10ef8478-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jphx9\" (UID: \"b47bfeba-2bc6-4c04-bd5a-a02b10ef8478\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598272 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69vgp\" (UniqueName: \"kubernetes.io/projected/c788e539-e984-4afb-ac5f-d216466ff704-kube-api-access-69vgp\") pod \"multus-admission-controller-857f4d67dd-h98fq\" (UID: \"c788e539-e984-4afb-ac5f-d216466ff704\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h98fq" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598292 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7nzl\" (UniqueName: \"kubernetes.io/projected/afa06ee0-ace4-4549-8bf4-2d02930cf2f5-kube-api-access-c7nzl\") pod \"kube-storage-version-migrator-operator-b67b599dd-nb6gw\" (UID: \"afa06ee0-ace4-4549-8bf4-2d02930cf2f5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598330 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c70614b5-a089-4e6e-8c59-06a01aaa2c54-node-bootstrap-token\") pod \"machine-config-server-8zrgd\" (UID: \"c70614b5-a089-4e6e-8c59-06a01aaa2c54\") " pod="openshift-machine-config-operator/machine-config-server-8zrgd" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598354 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qlg4\" (UniqueName: \"kubernetes.io/projected/bb2dc677-b493-45e9-a3a0-fc805b303a11-kube-api-access-4qlg4\") pod \"machine-config-controller-84d6567774-jpztt\" (UID: \"bb2dc677-b493-45e9-a3a0-fc805b303a11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598376 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ee69313-9ffd-4f06-aa3a-58037bd058de-metrics-tls\") pod \"dns-operator-744455d44c-s5dk6\" (UID: \"0ee69313-9ffd-4f06-aa3a-58037bd058de\") " pod="openshift-dns-operator/dns-operator-744455d44c-s5dk6" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598396 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6-apiservice-cert\") pod \"packageserver-d55dfcdfc-jtl86\" (UID: \"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598418 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d161354c-4c64-4c8f-a368-3462e7d3c5ac-metrics-certs\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.598438 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6-tmpfs\") pod \"packageserver-d55dfcdfc-jtl86\" (UID: \"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599232 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-mountpoint-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599273 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-config-volume\") pod \"collect-profiles-29340915-lj4bw\" (UID: \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599296 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab52f1c4-7d72-4d74-a16f-d49ed094b8e8-config\") pod \"kube-apiserver-operator-766d6c64bb-5v9kf\" (UID: \"ab52f1c4-7d72-4d74-a16f-d49ed094b8e8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599321 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/98994fb5-4f28-4381-9ae2-4c734bf79c5c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2kwtj\" (UID: \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599341 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/39778b3c-6ffc-417e-b9ff-50526b1fd867-profile-collector-cert\") pod \"catalog-operator-68c6474976-mk6kt\" (UID: \"39778b3c-6ffc-417e-b9ff-50526b1fd867\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599364 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d161354c-4c64-4c8f-a368-3462e7d3c5ac-default-certificate\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599553 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tr5r\" (UniqueName: \"kubernetes.io/projected/39778b3c-6ffc-417e-b9ff-50526b1fd867-kube-api-access-7tr5r\") pod \"catalog-operator-68c6474976-mk6kt\" (UID: \"39778b3c-6ffc-417e-b9ff-50526b1fd867\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599587 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98994fb5-4f28-4381-9ae2-4c734bf79c5c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2kwtj\" (UID: \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599610 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnpc4\" (UniqueName: \"kubernetes.io/projected/98994fb5-4f28-4381-9ae2-4c734bf79c5c-kube-api-access-rnpc4\") pod \"marketplace-operator-79b997595-2kwtj\" (UID: \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599630 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm7q8\" (UniqueName: \"kubernetes.io/projected/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-kube-api-access-wm7q8\") pod \"collect-profiles-29340915-lj4bw\" (UID: \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599650 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab52f1c4-7d72-4d74-a16f-d49ed094b8e8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5v9kf\" (UID: \"ab52f1c4-7d72-4d74-a16f-d49ed094b8e8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599670 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9a8a1bd5-206a-4154-886b-21f1b00906a6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfgxp\" (UID: \"9a8a1bd5-206a-4154-886b-21f1b00906a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599695 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afa06ee0-ace4-4549-8bf4-2d02930cf2f5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nb6gw\" (UID: \"afa06ee0-ace4-4549-8bf4-2d02930cf2f5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599714 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv85p\" (UniqueName: \"kubernetes.io/projected/ade216cb-9ef7-4dac-b35e-61dd3547b3c3-kube-api-access-fv85p\") pod \"machine-config-operator-74547568cd-m8jbn\" (UID: \"ade216cb-9ef7-4dac-b35e-61dd3547b3c3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.599735 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-secret-volume\") pod \"collect-profiles-29340915-lj4bw\" (UID: \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.600014 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx2x4\" (UniqueName: \"kubernetes.io/projected/49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6-kube-api-access-nx2x4\") pod \"packageserver-d55dfcdfc-jtl86\" (UID: \"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.602438 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98994fb5-4f28-4381-9ae2-4c734bf79c5c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2kwtj\" (UID: \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.603472 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cd5588a7-14c8-4a04-bccc-a22bc3edefaa-cert\") pod \"ingress-canary-2j7ss\" (UID: \"cd5588a7-14c8-4a04-bccc-a22bc3edefaa\") " pod="openshift-ingress-canary/ingress-canary-2j7ss" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.604283 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqc7b\" (UniqueName: \"kubernetes.io/projected/81c3c0f8-ad71-4a4b-8193-95051f481416-kube-api-access-bqc7b\") pod \"apiserver-76f77b778f-8f2dm\" (UID: \"81c3c0f8-ad71-4a4b-8193-95051f481416\") " pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.604464 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2-config\") pod \"service-ca-operator-777779d784-c25w8\" (UID: \"877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.604621 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6-tmpfs\") pod \"packageserver-d55dfcdfc-jtl86\" (UID: \"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.607548 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-config-volume\") pod \"collect-profiles-29340915-lj4bw\" (UID: \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.607621 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-mountpoint-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.607994 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c70614b5-a089-4e6e-8c59-06a01aaa2c54-certs\") pod \"machine-config-server-8zrgd\" (UID: \"c70614b5-a089-4e6e-8c59-06a01aaa2c54\") " pod="openshift-machine-config-operator/machine-config-server-8zrgd" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.608936 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab52f1c4-7d72-4d74-a16f-d49ed094b8e8-config\") pod \"kube-apiserver-operator-766d6c64bb-5v9kf\" (UID: \"ab52f1c4-7d72-4d74-a16f-d49ed094b8e8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.611667 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-socket-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.611800 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-registration-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.614251 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/39778b3c-6ffc-417e-b9ff-50526b1fd867-profile-collector-cert\") pod \"catalog-operator-68c6474976-mk6kt\" (UID: \"39778b3c-6ffc-417e-b9ff-50526b1fd867\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.616181 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d161354c-4c64-4c8f-a368-3462e7d3c5ac-default-certificate\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.616913 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" event={"ID":"7ff096c7-1a05-493a-9e19-8bec41303e6c","Type":"ContainerStarted","Data":"1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae"} Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.616951 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" event={"ID":"7ff096c7-1a05-493a-9e19-8bec41303e6c","Type":"ContainerStarted","Data":"205e4e3af51b98cfd46a1adf0e05e4332b649d95f8c3fe641d631defc95e1a37"} Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.618230 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619362 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.600070 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c70614b5-a089-4e6e-8c59-06a01aaa2c54-certs\") pod \"machine-config-server-8zrgd\" (UID: \"c70614b5-a089-4e6e-8c59-06a01aaa2c54\") " pod="openshift-machine-config-operator/machine-config-server-8zrgd" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619661 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fdb8607a-0a03-4e36-bf27-080330220d71-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9qkjs\" (UID: \"fdb8607a-0a03-4e36-bf27-080330220d71\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619693 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znqms\" (UniqueName: \"kubernetes.io/projected/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-kube-api-access-znqms\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619719 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a98ffecf-8060-4d01-8bfe-e1f7c8f04b46-signing-cabundle\") pod \"service-ca-9c57cc56f-9chhc\" (UID: \"a98ffecf-8060-4d01-8bfe-e1f7c8f04b46\") " pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619751 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619776 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv2cx\" (UniqueName: \"kubernetes.io/projected/d161354c-4c64-4c8f-a368-3462e7d3c5ac-kube-api-access-qv2cx\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619798 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd845cc3-5d83-4f84-9757-338b2eb40662-trusted-ca\") pod \"ingress-operator-5b745b69d9-dmdlz\" (UID: \"fd845cc3-5d83-4f84-9757-338b2eb40662\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619827 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d161354c-4c64-4c8f-a368-3462e7d3c5ac-service-ca-bundle\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619847 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa06ee0-ace4-4549-8bf4-2d02930cf2f5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nb6gw\" (UID: \"afa06ee0-ace4-4549-8bf4-2d02930cf2f5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619883 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkcfp\" (UniqueName: \"kubernetes.io/projected/fdb8607a-0a03-4e36-bf27-080330220d71-kube-api-access-nkcfp\") pod \"olm-operator-6b444d44fb-9qkjs\" (UID: \"fdb8607a-0a03-4e36-bf27-080330220d71\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619908 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fd845cc3-5d83-4f84-9757-338b2eb40662-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dmdlz\" (UID: \"fd845cc3-5d83-4f84-9757-338b2eb40662\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619932 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fdb8607a-0a03-4e36-bf27-080330220d71-srv-cert\") pod \"olm-operator-6b444d44fb-9qkjs\" (UID: \"fdb8607a-0a03-4e36-bf27-080330220d71\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619950 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-plugins-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619971 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a8a1bd5-206a-4154-886b-21f1b00906a6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfgxp\" (UID: \"9a8a1bd5-206a-4154-886b-21f1b00906a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.619994 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/39778b3c-6ffc-417e-b9ff-50526b1fd867-srv-cert\") pod \"catalog-operator-68c6474976-mk6kt\" (UID: \"39778b3c-6ffc-417e-b9ff-50526b1fd867\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620019 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-csi-data-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620043 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv2fz\" (UniqueName: \"kubernetes.io/projected/fd845cc3-5d83-4f84-9757-338b2eb40662-kube-api-access-vv2fz\") pod \"ingress-operator-5b745b69d9-dmdlz\" (UID: \"fd845cc3-5d83-4f84-9757-338b2eb40662\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620064 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2-serving-cert\") pod \"service-ca-operator-777779d784-c25w8\" (UID: \"877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620091 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g926\" (UniqueName: \"kubernetes.io/projected/c70614b5-a089-4e6e-8c59-06a01aaa2c54-kube-api-access-6g926\") pod \"machine-config-server-8zrgd\" (UID: \"c70614b5-a089-4e6e-8c59-06a01aaa2c54\") " pod="openshift-machine-config-operator/machine-config-server-8zrgd" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620112 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d161354c-4c64-4c8f-a368-3462e7d3c5ac-stats-auth\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620284 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ade216cb-9ef7-4dac-b35e-61dd3547b3c3-images\") pod \"machine-config-operator-74547568cd-m8jbn\" (UID: \"ade216cb-9ef7-4dac-b35e-61dd3547b3c3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620494 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdc28\" (UniqueName: \"kubernetes.io/projected/877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2-kube-api-access-rdc28\") pod \"service-ca-operator-777779d784-c25w8\" (UID: \"877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620531 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6-webhook-cert\") pod \"packageserver-d55dfcdfc-jtl86\" (UID: \"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620560 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ade216cb-9ef7-4dac-b35e-61dd3547b3c3-proxy-tls\") pod \"machine-config-operator-74547568cd-m8jbn\" (UID: \"ade216cb-9ef7-4dac-b35e-61dd3547b3c3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620583 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8952m\" (UniqueName: \"kubernetes.io/projected/a98ffecf-8060-4d01-8bfe-e1f7c8f04b46-kube-api-access-8952m\") pod \"service-ca-9c57cc56f-9chhc\" (UID: \"a98ffecf-8060-4d01-8bfe-e1f7c8f04b46\") " pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620606 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab52f1c4-7d72-4d74-a16f-d49ed094b8e8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5v9kf\" (UID: \"ab52f1c4-7d72-4d74-a16f-d49ed094b8e8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620631 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a8a1bd5-206a-4154-886b-21f1b00906a6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfgxp\" (UID: \"9a8a1bd5-206a-4154-886b-21f1b00906a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620664 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bb2dc677-b493-45e9-a3a0-fc805b303a11-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jpztt\" (UID: \"bb2dc677-b493-45e9-a3a0-fc805b303a11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.620755 4945 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-hxtjg container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.621607 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d161354c-4c64-4c8f-a368-3462e7d3c5ac-service-ca-bundle\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.621593 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" podUID="7ff096c7-1a05-493a-9e19-8bec41303e6c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 14 15:17:02 crc kubenswrapper[4945]: E1014 15:17:02.622320 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:03.122304044 +0000 UTC m=+153.106352412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.622393 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-plugins-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.623820 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a8a1bd5-206a-4154-886b-21f1b00906a6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfgxp\" (UID: \"9a8a1bd5-206a-4154-886b-21f1b00906a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.623961 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-csi-data-dir\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.627657 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa06ee0-ace4-4549-8bf4-2d02930cf2f5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nb6gw\" (UID: \"afa06ee0-ace4-4549-8bf4-2d02930cf2f5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.631086 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab52f1c4-7d72-4d74-a16f-d49ed094b8e8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5v9kf\" (UID: \"ab52f1c4-7d72-4d74-a16f-d49ed094b8e8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.632667 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a98ffecf-8060-4d01-8bfe-e1f7c8f04b46-signing-cabundle\") pod \"service-ca-9c57cc56f-9chhc\" (UID: \"a98ffecf-8060-4d01-8bfe-e1f7c8f04b46\") " pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.636318 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ee69313-9ffd-4f06-aa3a-58037bd058de-metrics-tls\") pod \"dns-operator-744455d44c-s5dk6\" (UID: \"0ee69313-9ffd-4f06-aa3a-58037bd058de\") " pod="openshift-dns-operator/dns-operator-744455d44c-s5dk6" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.638973 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd845cc3-5d83-4f84-9757-338b2eb40662-trusted-ca\") pod \"ingress-operator-5b745b69d9-dmdlz\" (UID: \"fd845cc3-5d83-4f84-9757-338b2eb40662\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.639452 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fd845cc3-5d83-4f84-9757-338b2eb40662-metrics-tls\") pod \"ingress-operator-5b745b69d9-dmdlz\" (UID: \"fd845cc3-5d83-4f84-9757-338b2eb40662\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.640816 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ade216cb-9ef7-4dac-b35e-61dd3547b3c3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-m8jbn\" (UID: \"ade216cb-9ef7-4dac-b35e-61dd3547b3c3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.643624 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rm8t5"] Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.644461 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/748d0e0e-3657-4ad3-8918-c0d0773e06e6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8rkjt\" (UID: \"748d0e0e-3657-4ad3-8918-c0d0773e06e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.645774 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6-apiservice-cert\") pod \"packageserver-d55dfcdfc-jtl86\" (UID: \"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.647222 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a8a1bd5-206a-4154-886b-21f1b00906a6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfgxp\" (UID: \"9a8a1bd5-206a-4154-886b-21f1b00906a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.647249 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a98ffecf-8060-4d01-8bfe-e1f7c8f04b46-signing-key\") pod \"service-ca-9c57cc56f-9chhc\" (UID: \"a98ffecf-8060-4d01-8bfe-e1f7c8f04b46\") " pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.647715 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d161354c-4c64-4c8f-a368-3462e7d3c5ac-metrics-certs\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.647795 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afa06ee0-ace4-4549-8bf4-2d02930cf2f5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nb6gw\" (UID: \"afa06ee0-ace4-4549-8bf4-2d02930cf2f5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.649381 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bb2dc677-b493-45e9-a3a0-fc805b303a11-proxy-tls\") pod \"machine-config-controller-84d6567774-jpztt\" (UID: \"bb2dc677-b493-45e9-a3a0-fc805b303a11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.650016 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fdb8607a-0a03-4e36-bf27-080330220d71-srv-cert\") pod \"olm-operator-6b444d44fb-9qkjs\" (UID: \"fdb8607a-0a03-4e36-bf27-080330220d71\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.650276 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fdb8607a-0a03-4e36-bf27-080330220d71-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9qkjs\" (UID: \"fdb8607a-0a03-4e36-bf27-080330220d71\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.652719 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/98994fb5-4f28-4381-9ae2-4c734bf79c5c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2kwtj\" (UID: \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.653025 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgjgp\" (UniqueName: \"kubernetes.io/projected/06ecdccd-2560-48ea-a7f3-87e3e75488a0-kube-api-access-mgjgp\") pod \"console-f9d7485db-ccxsw\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.654027 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d161354c-4c64-4c8f-a368-3462e7d3c5ac-stats-auth\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.654304 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-secret-volume\") pod \"collect-profiles-29340915-lj4bw\" (UID: \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.654818 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c788e539-e984-4afb-ac5f-d216466ff704-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-h98fq\" (UID: \"c788e539-e984-4afb-ac5f-d216466ff704\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h98fq" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.654962 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b47bfeba-2bc6-4c04-bd5a-a02b10ef8478-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jphx9\" (UID: \"b47bfeba-2bc6-4c04-bd5a-a02b10ef8478\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.655386 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/39778b3c-6ffc-417e-b9ff-50526b1fd867-srv-cert\") pod \"catalog-operator-68c6474976-mk6kt\" (UID: \"39778b3c-6ffc-417e-b9ff-50526b1fd867\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.657465 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ade216cb-9ef7-4dac-b35e-61dd3547b3c3-proxy-tls\") pod \"machine-config-operator-74547568cd-m8jbn\" (UID: \"ade216cb-9ef7-4dac-b35e-61dd3547b3c3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.658185 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2-serving-cert\") pod \"service-ca-operator-777779d784-c25w8\" (UID: \"877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.659630 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c70614b5-a089-4e6e-8c59-06a01aaa2c54-node-bootstrap-token\") pod \"machine-config-server-8zrgd\" (UID: \"c70614b5-a089-4e6e-8c59-06a01aaa2c54\") " pod="openshift-machine-config-operator/machine-config-server-8zrgd" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.662141 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6-webhook-cert\") pod \"packageserver-d55dfcdfc-jtl86\" (UID: \"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.662220 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c26cc\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-kube-api-access-c26cc\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.666751 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6cv87"] Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.676698 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khq4r\" (UniqueName: \"kubernetes.io/projected/748d0e0e-3657-4ad3-8918-c0d0773e06e6-kube-api-access-khq4r\") pod \"package-server-manager-789f6589d5-8rkjt\" (UID: \"748d0e0e-3657-4ad3-8918-c0d0773e06e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.689296 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" Oct 14 15:17:02 crc kubenswrapper[4945]: W1014 15:17:02.700514 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ce14cdf_3467_4035_bb1b_23a621ed6a93.slice/crio-4de49fd054dc471b86e835baaf120cff00ae378ba79d710b6455c02d03a2ab32 WatchSource:0}: Error finding container 4de49fd054dc471b86e835baaf120cff00ae378ba79d710b6455c02d03a2ab32: Status 404 returned error can't find the container with id 4de49fd054dc471b86e835baaf120cff00ae378ba79d710b6455c02d03a2ab32 Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.702838 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krgsh\" (UniqueName: \"kubernetes.io/projected/cd5588a7-14c8-4a04-bccc-a22bc3edefaa-kube-api-access-krgsh\") pod \"ingress-canary-2j7ss\" (UID: \"cd5588a7-14c8-4a04-bccc-a22bc3edefaa\") " pod="openshift-ingress-canary/ingress-canary-2j7ss" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.720107 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx2x4\" (UniqueName: \"kubernetes.io/projected/49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6-kube-api-access-nx2x4\") pod \"packageserver-d55dfcdfc-jtl86\" (UID: \"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.729320 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:02 crc kubenswrapper[4945]: E1014 15:17:02.729983 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:03.229957024 +0000 UTC m=+153.214005392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.743597 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.749727 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th2mh\" (UniqueName: \"kubernetes.io/projected/0ee69313-9ffd-4f06-aa3a-58037bd058de-kube-api-access-th2mh\") pod \"dns-operator-744455d44c-s5dk6\" (UID: \"0ee69313-9ffd-4f06-aa3a-58037bd058de\") " pod="openshift-dns-operator/dns-operator-744455d44c-s5dk6" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.751233 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.768431 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm7q8\" (UniqueName: \"kubernetes.io/projected/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-kube-api-access-wm7q8\") pod \"collect-profiles-29340915-lj4bw\" (UID: \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.782951 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tr5r\" (UniqueName: \"kubernetes.io/projected/39778b3c-6ffc-417e-b9ff-50526b1fd867-kube-api-access-7tr5r\") pod \"catalog-operator-68c6474976-mk6kt\" (UID: \"39778b3c-6ffc-417e-b9ff-50526b1fd867\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.811563 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnpc4\" (UniqueName: \"kubernetes.io/projected/98994fb5-4f28-4381-9ae2-4c734bf79c5c-kube-api-access-rnpc4\") pod \"marketplace-operator-79b997595-2kwtj\" (UID: \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.820045 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9a8a1bd5-206a-4154-886b-21f1b00906a6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfgxp\" (UID: \"9a8a1bd5-206a-4154-886b-21f1b00906a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.820268 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.831210 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:02 crc kubenswrapper[4945]: E1014 15:17:02.831621 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:03.331605605 +0000 UTC m=+153.315653973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.837241 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-s5dk6" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.847980 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69vgp\" (UniqueName: \"kubernetes.io/projected/c788e539-e984-4afb-ac5f-d216466ff704-kube-api-access-69vgp\") pod \"multus-admission-controller-857f4d67dd-h98fq\" (UID: \"c788e539-e984-4afb-ac5f-d216466ff704\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h98fq" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.862961 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4md4\" (UniqueName: \"kubernetes.io/projected/d5913ac3-e4f4-4397-bc8c-3fd7ab4fee4d-kube-api-access-w4md4\") pod \"migrator-59844c95c7-45pg9\" (UID: \"d5913ac3-e4f4-4397-bc8c-3fd7ab4fee4d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-45pg9" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.872059 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2j7ss" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.891034 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.901796 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhzng\" (UniqueName: \"kubernetes.io/projected/b47bfeba-2bc6-4c04-bd5a-a02b10ef8478-kube-api-access-fhzng\") pod \"control-plane-machine-set-operator-78cbb6b69f-jphx9\" (UID: \"b47bfeba-2bc6-4c04-bd5a-a02b10ef8478\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.917198 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv85p\" (UniqueName: \"kubernetes.io/projected/ade216cb-9ef7-4dac-b35e-61dd3547b3c3-kube-api-access-fv85p\") pod \"machine-config-operator-74547568cd-m8jbn\" (UID: \"ade216cb-9ef7-4dac-b35e-61dd3547b3c3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.925808 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7nzl\" (UniqueName: \"kubernetes.io/projected/afa06ee0-ace4-4549-8bf4-2d02930cf2f5-kube-api-access-c7nzl\") pod \"kube-storage-version-migrator-operator-b67b599dd-nb6gw\" (UID: \"afa06ee0-ace4-4549-8bf4-2d02930cf2f5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.932254 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:02 crc kubenswrapper[4945]: E1014 15:17:02.932695 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:03.43267869 +0000 UTC m=+153.416727058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.938996 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qlg4\" (UniqueName: \"kubernetes.io/projected/bb2dc677-b493-45e9-a3a0-fc805b303a11-kube-api-access-4qlg4\") pod \"machine-config-controller-84d6567774-jpztt\" (UID: \"bb2dc677-b493-45e9-a3a0-fc805b303a11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.958310 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znqms\" (UniqueName: \"kubernetes.io/projected/c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc-kube-api-access-znqms\") pod \"csi-hostpathplugin-tgs6r\" (UID: \"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc\") " pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:02 crc kubenswrapper[4945]: I1014 15:17:02.979137 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv2cx\" (UniqueName: \"kubernetes.io/projected/d161354c-4c64-4c8f-a368-3462e7d3c5ac-kube-api-access-qv2cx\") pod \"router-default-5444994796-6j5pp\" (UID: \"d161354c-4c64-4c8f-a368-3462e7d3c5ac\") " pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.004636 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.010688 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkcfp\" (UniqueName: \"kubernetes.io/projected/fdb8607a-0a03-4e36-bf27-080330220d71-kube-api-access-nkcfp\") pod \"olm-operator-6b444d44fb-9qkjs\" (UID: \"fdb8607a-0a03-4e36-bf27-080330220d71\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.012495 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.026316 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fd845cc3-5d83-4f84-9757-338b2eb40662-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dmdlz\" (UID: \"fd845cc3-5d83-4f84-9757-338b2eb40662\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.029775 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.035770 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:03 crc kubenswrapper[4945]: E1014 15:17:03.036205 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:03.536188554 +0000 UTC m=+153.520236922 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.037483 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-45pg9" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.040115 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv2fz\" (UniqueName: \"kubernetes.io/projected/fd845cc3-5d83-4f84-9757-338b2eb40662-kube-api-access-vv2fz\") pod \"ingress-operator-5b745b69d9-dmdlz\" (UID: \"fd845cc3-5d83-4f84-9757-338b2eb40662\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.058894 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.062283 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g926\" (UniqueName: \"kubernetes.io/projected/c70614b5-a089-4e6e-8c59-06a01aaa2c54-kube-api-access-6g926\") pod \"machine-config-server-8zrgd\" (UID: \"c70614b5-a089-4e6e-8c59-06a01aaa2c54\") " pod="openshift-machine-config-operator/machine-config-server-8zrgd" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.070165 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.082069 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.084387 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdc28\" (UniqueName: \"kubernetes.io/projected/877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2-kube-api-access-rdc28\") pod \"service-ca-operator-777779d784-c25w8\" (UID: \"877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.089238 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.100439 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8952m\" (UniqueName: \"kubernetes.io/projected/a98ffecf-8060-4d01-8bfe-e1f7c8f04b46-kube-api-access-8952m\") pod \"service-ca-9c57cc56f-9chhc\" (UID: \"a98ffecf-8060-4d01-8bfe-e1f7c8f04b46\") " pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.104091 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.112962 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.127456 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab52f1c4-7d72-4d74-a16f-d49ed094b8e8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5v9kf\" (UID: \"ab52f1c4-7d72-4d74-a16f-d49ed094b8e8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.127696 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-h98fq" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.137708 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj"] Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.138313 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:03 crc kubenswrapper[4945]: E1014 15:17:03.138684 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:03.638669589 +0000 UTC m=+153.622717957 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.143775 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8zrgd" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.150444 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc"] Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.152098 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52"] Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.169115 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.239569 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:03 crc kubenswrapper[4945]: E1014 15:17:03.239988 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:03.73997451 +0000 UTC m=+153.724022878 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.295577 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.321680 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.340337 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:03 crc kubenswrapper[4945]: E1014 15:17:03.340675 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:03.840657875 +0000 UTC m=+153.824706243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.372781 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.384413 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf"] Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.388978 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-frj5t"] Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.396562 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.441844 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:03 crc kubenswrapper[4945]: E1014 15:17:03.442155 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:03.942141381 +0000 UTC m=+153.926189739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.545191 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:03 crc kubenswrapper[4945]: E1014 15:17:03.545504 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:04.045445269 +0000 UTC m=+154.029493637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.545668 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:03 crc kubenswrapper[4945]: E1014 15:17:03.545967 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:04.045956054 +0000 UTC m=+154.030004422 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.585360 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86"] Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.631977 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8f2dm"] Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.645564 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9zzcq"] Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.646340 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:03 crc kubenswrapper[4945]: E1014 15:17:03.646397 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:04.14638254 +0000 UTC m=+154.130430898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.646983 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:03 crc kubenswrapper[4945]: E1014 15:17:03.647419 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:04.147403579 +0000 UTC m=+154.131451947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.660457 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-6j5pp" event={"ID":"d161354c-4c64-4c8f-a368-3462e7d3c5ac","Type":"ContainerStarted","Data":"6ea87b43263271fad0809b58f1f6182e0e48e42ee3ea79dc5698495969c8b796"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.667387 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" event={"ID":"638d7074-0fbb-4086-9c1d-37036bee9034","Type":"ContainerStarted","Data":"5a595d710102a4f615b65dcd072ff0118ac94902b3760ffffd6faa3c9210ed24"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.697925 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8zrgd" event={"ID":"c70614b5-a089-4e6e-8c59-06a01aaa2c54","Type":"ContainerStarted","Data":"26e2bb2a463c84d22ed3a1130bdca1423af40301a8feb9fece9311bf931bf75f"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.705773 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6cv87" event={"ID":"58804eb4-1016-427d-bcf6-fcfa9e8864df","Type":"ContainerStarted","Data":"8209f71bc473bda9918318cf5838a5f19302b407f74747fec4d016d0154538a9"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.705810 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6cv87" event={"ID":"58804eb4-1016-427d-bcf6-fcfa9e8864df","Type":"ContainerStarted","Data":"26a8999aa26f353b81d03696b9fc0cb74f7bf0e8eb8cf7ff2da0245f8047ecc4"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.708185 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6cv87" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.709357 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cv87 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.712336 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" event={"ID":"b683cb73-6094-43f6-b9f3-9f89691405c1","Type":"ContainerStarted","Data":"01bbfc2836126f7f465d177e5db870fc6fbd6af69056f17f7d9a93156761ecd8"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.714845 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cv87" podUID="58804eb4-1016-427d-bcf6-fcfa9e8864df" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.725186 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc" event={"ID":"82aa3bda-5c5a-4d70-826e-e0a5d358e3f2","Type":"ContainerStarted","Data":"cad5a298f1b5737310abe7d96bd36b57dce9171148cd567d796d1e98f66db3fc"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.748004 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:03 crc kubenswrapper[4945]: E1014 15:17:03.749366 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:04.249349099 +0000 UTC m=+154.233397467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.761006 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" event={"ID":"388c7278-fae0-4de0-b646-49e46f88c517","Type":"ContainerStarted","Data":"d6eeba569d1eb0b8ee0e9a121f6b108487f847357854651e541d5bbea8055a56"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.764793 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" event={"ID":"b1593874-4de8-4e33-aadb-0dbdf30e7262","Type":"ContainerStarted","Data":"fd828869b035631bf7e487d146ee368877ea861d62f8c721a7abb6a069fa08de"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.791148 4945 generic.go:334] "Generic (PLEG): container finished" podID="50be8f46-65da-4733-bb6d-b4dc67c03b60" containerID="bd91afaa66e4ec8bbd520a4e7f0b996e118c047ab347b788194a708ca34d2d5d" exitCode=0 Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.791265 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" event={"ID":"50be8f46-65da-4733-bb6d-b4dc67c03b60","Type":"ContainerDied","Data":"bd91afaa66e4ec8bbd520a4e7f0b996e118c047ab347b788194a708ca34d2d5d"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.791295 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" event={"ID":"50be8f46-65da-4733-bb6d-b4dc67c03b60","Type":"ContainerStarted","Data":"536a1042b80c4ba8724d54a51a70477866d3e0d53f4bef6a75d9a8294c4b2834"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.795920 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" event={"ID":"932ab187-513c-4d0f-9c6f-f986ac444f78","Type":"ContainerStarted","Data":"d905493f2e9d74dd4fe13f1b81f000f5579d6b67136a4f9efe314d5d22ea074f"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.795947 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" event={"ID":"932ab187-513c-4d0f-9c6f-f986ac444f78","Type":"ContainerStarted","Data":"c311e90afae3350c03ec46ffaf276c4822fef9ab58a23250ed858b71002a37de"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.799909 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hvp57" event={"ID":"0ce14cdf-3467-4035-bb1b-23a621ed6a93","Type":"ContainerStarted","Data":"0c685d9dc5e0f8466c49faba1c8b352ac60bdeab47a972b1a1aa59048e115535"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.799939 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hvp57" event={"ID":"0ce14cdf-3467-4035-bb1b-23a621ed6a93","Type":"ContainerStarted","Data":"4de49fd054dc471b86e835baaf120cff00ae378ba79d710b6455c02d03a2ab32"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.801001 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.801763 4945 patch_prober.go:28] interesting pod/console-operator-58897d9998-hvp57 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.801789 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" event={"ID":"d55256a5-30da-4fd3-beef-2d3c785ffaa4","Type":"ContainerStarted","Data":"39afd9de149eae3e323e99d1c15b59ae2dc9afdd036b44f17a05eebed7b03d22"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.801796 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hvp57" podUID="0ce14cdf-3467-4035-bb1b-23a621ed6a93" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.801808 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" event={"ID":"d55256a5-30da-4fd3-beef-2d3c785ffaa4","Type":"ContainerStarted","Data":"9150e2e73925bbc9bf55f94c41619060e3fde93972852c98b61da400cedc6a7f"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.808593 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" event={"ID":"d8cdf424-f208-4e85-a001-0ffc3a1b27c2","Type":"ContainerStarted","Data":"74e908d1b4ff02a86253009f6d35f33db4f4555a2a6dc3bac38d2e2774824f2d"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.810033 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" event={"ID":"58633adf-8733-4b72-a9da-78e278ebe5fd","Type":"ContainerStarted","Data":"8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.813318 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.816220 4945 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-glv7n container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.816266 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" podUID="58633adf-8733-4b72-a9da-78e278ebe5fd" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.830510 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-l72k8" event={"ID":"2c3400fe-3438-4116-b141-3281b0fa2d75","Type":"ContainerStarted","Data":"5c5d2541855cd909b62c25b8ad476c4a0f0522ab3b874f06080dc4a3b07e96f0"} Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.830571 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-l72k8" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.834514 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.850493 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:03 crc kubenswrapper[4945]: E1014 15:17:03.850823 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:04.350805625 +0000 UTC m=+154.334853993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:03 crc kubenswrapper[4945]: W1014 15:17:03.921102 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd094c180_e6dd_4064_aea1_4c04b43e4963.slice/crio-3b88c808254ceeb048e67aa5549253f449702b2188e3500c514336d6fbc575b2 WatchSource:0}: Error finding container 3b88c808254ceeb048e67aa5549253f449702b2188e3500c514336d6fbc575b2: Status 404 returned error can't find the container with id 3b88c808254ceeb048e67aa5549253f449702b2188e3500c514336d6fbc575b2 Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.951220 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:03 crc kubenswrapper[4945]: E1014 15:17:03.952292 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:04.452265161 +0000 UTC m=+154.436313569 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:03 crc kubenswrapper[4945]: I1014 15:17:03.954760 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkb5l" podStartSLOduration=133.954747101 podStartE2EDuration="2m13.954747101s" podCreationTimestamp="2025-10-14 15:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:03.95151447 +0000 UTC m=+153.935562848" watchObservedRunningTime="2025-10-14 15:17:03.954747101 +0000 UTC m=+153.938795469" Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.057493 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:04 crc kubenswrapper[4945]: E1014 15:17:04.058042 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:04.558031338 +0000 UTC m=+154.542079706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.088401 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.110928 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.121347 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp"] Oct 14 15:17:04 crc kubenswrapper[4945]: W1014 15:17:04.151808 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39778b3c_6ffc_417e_b9ff_50526b1fd867.slice/crio-05ea3a0c58e6b4bd5d6026304518e812cd6e22b9aeeb1d22f545dca49be33035 WatchSource:0}: Error finding container 05ea3a0c58e6b4bd5d6026304518e812cd6e22b9aeeb1d22f545dca49be33035: Status 404 returned error can't find the container with id 05ea3a0c58e6b4bd5d6026304518e812cd6e22b9aeeb1d22f545dca49be33035 Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.158520 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:04 crc kubenswrapper[4945]: E1014 15:17:04.158794 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:04.658778024 +0000 UTC m=+154.642826392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.162576 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2j7ss"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.174860 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-s5dk6"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.186815 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ccxsw"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.186912 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw"] Oct 14 15:17:04 crc kubenswrapper[4945]: W1014 15:17:04.199741 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafa06ee0_ace4_4549_8bf4_2d02930cf2f5.slice/crio-7adb722db7eda62fccb0a0a3f708fa4eec067bbb038c87c4edb20fa4bc2f8a99 WatchSource:0}: Error finding container 7adb722db7eda62fccb0a0a3f708fa4eec067bbb038c87c4edb20fa4bc2f8a99: Status 404 returned error can't find the container with id 7adb722db7eda62fccb0a0a3f708fa4eec067bbb038c87c4edb20fa4bc2f8a99 Oct 14 15:17:04 crc kubenswrapper[4945]: W1014 15:17:04.207509 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06ecdccd_2560_48ea_a7f3_87e3e75488a0.slice/crio-354b14026fcbff4bf145f5c84b4f72443811c15403dc6c5671c2b01813577872 WatchSource:0}: Error finding container 354b14026fcbff4bf145f5c84b4f72443811c15403dc6c5671c2b01813577872: Status 404 returned error can't find the container with id 354b14026fcbff4bf145f5c84b4f72443811c15403dc6c5671c2b01813577872 Oct 14 15:17:04 crc kubenswrapper[4945]: W1014 15:17:04.230828 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd5588a7_14c8_4a04_bccc_a22bc3edefaa.slice/crio-a0357389c41ba3e2f2ed9722591e25e1f1358bd4a893e64bf66af7dfcf56091d WatchSource:0}: Error finding container a0357389c41ba3e2f2ed9722591e25e1f1358bd4a893e64bf66af7dfcf56091d: Status 404 returned error can't find the container with id a0357389c41ba3e2f2ed9722591e25e1f1358bd4a893e64bf66af7dfcf56091d Oct 14 15:17:04 crc kubenswrapper[4945]: E1014 15:17:04.260124 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:04.760113716 +0000 UTC m=+154.744162084 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.260412 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.284174 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.293670 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.325259 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.353360 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-h98fq"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.362410 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:04 crc kubenswrapper[4945]: E1014 15:17:04.362768 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:04.862750995 +0000 UTC m=+154.846799363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.364535 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.413694 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs"] Oct 14 15:17:04 crc kubenswrapper[4945]: W1014 15:17:04.424000 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc788e539_e984_4afb_ac5f_d216466ff704.slice/crio-3b44f2e0df279b28b547b350d440a9484b9c2a7bac1692ca8d6aad3c72d90b38 WatchSource:0}: Error finding container 3b44f2e0df279b28b547b350d440a9484b9c2a7bac1692ca8d6aad3c72d90b38: Status 404 returned error can't find the container with id 3b44f2e0df279b28b547b350d440a9484b9c2a7bac1692ca8d6aad3c72d90b38 Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.426826 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.426892 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2kwtj"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.428261 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-45pg9"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.434939 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.444669 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-tgs6r"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.466059 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:04 crc kubenswrapper[4945]: E1014 15:17:04.466315 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:04.9663055 +0000 UTC m=+154.950353868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.540388 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" podStartSLOduration=133.540369615 podStartE2EDuration="2m13.540369615s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:04.540269742 +0000 UTC m=+154.524318130" watchObservedRunningTime="2025-10-14 15:17:04.540369615 +0000 UTC m=+154.524417983" Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.573557 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:04 crc kubenswrapper[4945]: E1014 15:17:04.573858 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:05.073844358 +0000 UTC m=+155.057892726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:04 crc kubenswrapper[4945]: W1014 15:17:04.617030 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98994fb5_4f28_4381_9ae2_4c734bf79c5c.slice/crio-03ed83a4e5fbb9e657f904694242a6c57c79a96d47b4a3094938e93242f5bbf1 WatchSource:0}: Error finding container 03ed83a4e5fbb9e657f904694242a6c57c79a96d47b4a3094938e93242f5bbf1: Status 404 returned error can't find the container with id 03ed83a4e5fbb9e657f904694242a6c57c79a96d47b4a3094938e93242f5bbf1 Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.630216 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.650016 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9chhc"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.670644 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-c25w8"] Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.675796 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:04 crc kubenswrapper[4945]: E1014 15:17:04.676217 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:05.176206149 +0000 UTC m=+155.160254507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:04 crc kubenswrapper[4945]: W1014 15:17:04.730510 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab52f1c4_7d72_4d74_a16f_d49ed094b8e8.slice/crio-f86c6a07a77f9433bcfc795d20db72706883f14e1b6202be36f1a5fa29865ec2 WatchSource:0}: Error finding container f86c6a07a77f9433bcfc795d20db72706883f14e1b6202be36f1a5fa29865ec2: Status 404 returned error can't find the container with id f86c6a07a77f9433bcfc795d20db72706883f14e1b6202be36f1a5fa29865ec2 Oct 14 15:17:04 crc kubenswrapper[4945]: W1014 15:17:04.730940 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda98ffecf_8060_4d01_8bfe_e1f7c8f04b46.slice/crio-1b34e9bb3ab7766107551266191a462b7d6732c64240fdfdf3a54e413309b13d WatchSource:0}: Error finding container 1b34e9bb3ab7766107551266191a462b7d6732c64240fdfdf3a54e413309b13d: Status 404 returned error can't find the container with id 1b34e9bb3ab7766107551266191a462b7d6732c64240fdfdf3a54e413309b13d Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.776995 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:04 crc kubenswrapper[4945]: E1014 15:17:04.777137 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:05.277110989 +0000 UTC m=+155.261159357 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.777322 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:04 crc kubenswrapper[4945]: E1014 15:17:04.777631 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:05.277619064 +0000 UTC m=+155.261667432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.785629 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-fkbmp" podStartSLOduration=133.785611589 podStartE2EDuration="2m13.785611589s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:04.747586818 +0000 UTC m=+154.731635196" watchObservedRunningTime="2025-10-14 15:17:04.785611589 +0000 UTC m=+154.769659957" Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.785846 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78cw8" podStartSLOduration=134.785842895 podStartE2EDuration="2m14.785842895s" podCreationTimestamp="2025-10-14 15:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:04.784867508 +0000 UTC m=+154.768915876" watchObservedRunningTime="2025-10-14 15:17:04.785842895 +0000 UTC m=+154.769891263" Oct 14 15:17:04 crc kubenswrapper[4945]: W1014 15:17:04.873994 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod877b02e6_9caf_4066_bc2c_4d4d3cf8b7e2.slice/crio-efe60b49c6edb559c8f1f34f7cacecbf8c1b88f2e6468f1416b391957ad17f53 WatchSource:0}: Error finding container efe60b49c6edb559c8f1f34f7cacecbf8c1b88f2e6468f1416b391957ad17f53: Status 404 returned error can't find the container with id efe60b49c6edb559c8f1f34f7cacecbf8c1b88f2e6468f1416b391957ad17f53 Oct 14 15:17:04 crc kubenswrapper[4945]: E1014 15:17:04.880003 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:05.379980185 +0000 UTC m=+155.364028553 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.879913 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.880945 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:04 crc kubenswrapper[4945]: E1014 15:17:04.881801 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:05.381777966 +0000 UTC m=+155.365826334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.882619 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-rm8t5" podStartSLOduration=133.882599119 podStartE2EDuration="2m13.882599119s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:04.879359807 +0000 UTC m=+154.863408175" watchObservedRunningTime="2025-10-14 15:17:04.882599119 +0000 UTC m=+154.866647487" Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.890167 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-hvp57" podStartSLOduration=133.890148261 podStartE2EDuration="2m13.890148261s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:04.835782981 +0000 UTC m=+154.819831389" watchObservedRunningTime="2025-10-14 15:17:04.890148261 +0000 UTC m=+154.874196629" Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.919417 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" podStartSLOduration=133.919402775 podStartE2EDuration="2m13.919402775s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:04.918920121 +0000 UTC m=+154.902968489" watchObservedRunningTime="2025-10-14 15:17:04.919402775 +0000 UTC m=+154.903451143" Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.926259 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" event={"ID":"98994fb5-4f28-4381-9ae2-4c734bf79c5c","Type":"ContainerStarted","Data":"03ed83a4e5fbb9e657f904694242a6c57c79a96d47b4a3094938e93242f5bbf1"} Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.940936 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" event={"ID":"a98ffecf-8060-4d01-8bfe-e1f7c8f04b46","Type":"ContainerStarted","Data":"1b34e9bb3ab7766107551266191a462b7d6732c64240fdfdf3a54e413309b13d"} Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.946554 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-6j5pp" event={"ID":"d161354c-4c64-4c8f-a368-3462e7d3c5ac","Type":"ContainerStarted","Data":"73f48e50fbe359718fae8300483c91b5fad22a8e0a6f1e5fe8f13e10a635682d"} Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.950309 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-45pg9" event={"ID":"d5913ac3-e4f4-4397-bc8c-3fd7ab4fee4d","Type":"ContainerStarted","Data":"3c543b4d2135295f1fd491a54001b3ec6d6c4cfc4f9f0de7c7762fc07ca1b863"} Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.965800 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-h98fq" event={"ID":"c788e539-e984-4afb-ac5f-d216466ff704","Type":"ContainerStarted","Data":"3b44f2e0df279b28b547b350d440a9484b9c2a7bac1692ca8d6aad3c72d90b38"} Oct 14 15:17:04 crc kubenswrapper[4945]: I1014 15:17:04.991824 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:04 crc kubenswrapper[4945]: E1014 15:17:04.992247 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:05.492234705 +0000 UTC m=+155.476283073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.009150 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" event={"ID":"748d0e0e-3657-4ad3-8918-c0d0773e06e6","Type":"ContainerStarted","Data":"efc36948279dfdf983ed2af0dcffa4fd52de4452f6be6a7b6c498a71412bffc2"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.009196 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" event={"ID":"748d0e0e-3657-4ad3-8918-c0d0773e06e6","Type":"ContainerStarted","Data":"8b431162a5a7018a18b2702b3f875655cac0dad3d45e986ff33ec9cd88a0c255"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.017126 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-6cv87" podStartSLOduration=134.017109905 podStartE2EDuration="2m14.017109905s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:05.016855358 +0000 UTC m=+155.000903726" watchObservedRunningTime="2025-10-14 15:17:05.017109905 +0000 UTC m=+155.001158273" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.017334 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" event={"ID":"50be8f46-65da-4733-bb6d-b4dc67c03b60","Type":"ContainerStarted","Data":"bd4f48a52a6889cd07e8d0e06c097824e593a9b3b266a0c3e8ea8d3503280158"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.017911 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.018698 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-l72k8" podStartSLOduration=6.018688409 podStartE2EDuration="6.018688409s" podCreationTimestamp="2025-10-14 15:16:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:04.998371057 +0000 UTC m=+154.982419425" watchObservedRunningTime="2025-10-14 15:17:05.018688409 +0000 UTC m=+155.002736777" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.027943 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9" event={"ID":"b47bfeba-2bc6-4c04-bd5a-a02b10ef8478","Type":"ContainerStarted","Data":"82b37c0b5ee0ea8e5ed4516267a500f78abce9b08edae22d4435d7edd327d1cd"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.038465 4945 generic.go:334] "Generic (PLEG): container finished" podID="81c3c0f8-ad71-4a4b-8193-95051f481416" containerID="9430bc22139fab157aa6effc42362f345d8406d1437f5e476a5dd8398bffbd5a" exitCode=0 Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.038733 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" event={"ID":"81c3c0f8-ad71-4a4b-8193-95051f481416","Type":"ContainerDied","Data":"9430bc22139fab157aa6effc42362f345d8406d1437f5e476a5dd8398bffbd5a"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.038858 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" event={"ID":"81c3c0f8-ad71-4a4b-8193-95051f481416","Type":"ContainerStarted","Data":"6fb01ad9991c10df6116b12fb56394e8d25a51bd7c046931088bfcec82d824d1"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.063503 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.076402 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:05 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:05 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:05 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.076455 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.084178 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" event={"ID":"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc","Type":"ContainerStarted","Data":"15e09430375cf69eeb293f8d8ef97954de9f191244f0daa03f7150a118a20a15"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.089899 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" event={"ID":"3ec13203-2a26-4de1-b02f-30c58d8ab5a1","Type":"ContainerStarted","Data":"87556adef92a50d930116aceed4401b8295109ea34de9993164f10e7e97e090e"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.093241 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" event={"ID":"b1593874-4de8-4e33-aadb-0dbdf30e7262","Type":"ContainerStarted","Data":"d1fbb37930b9eda2cdc24f57e5089580e4fdfbff85cc493db07019b848918339"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.093528 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.093970 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:05.593957728 +0000 UTC m=+155.578006096 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.101910 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" event={"ID":"d094c180-e6dd-4064-aea1-4c04b43e4963","Type":"ContainerStarted","Data":"c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.101950 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" event={"ID":"d094c180-e6dd-4064-aea1-4c04b43e4963","Type":"ContainerStarted","Data":"3b88c808254ceeb048e67aa5549253f449702b2188e3500c514336d6fbc575b2"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.102264 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.125743 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5cj52" podStartSLOduration=134.125727152 podStartE2EDuration="2m14.125727152s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:05.073938235 +0000 UTC m=+155.057986603" watchObservedRunningTime="2025-10-14 15:17:05.125727152 +0000 UTC m=+155.109775520" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.127241 4945 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-9zzcq container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.20:6443/healthz\": dial tcp 10.217.0.20:6443: connect: connection refused" start-of-body= Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.127275 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" podUID="d094c180-e6dd-4064-aea1-4c04b43e4963" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.20:6443/healthz\": dial tcp 10.217.0.20:6443: connect: connection refused" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.130351 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" event={"ID":"fdb8607a-0a03-4e36-bf27-080330220d71","Type":"ContainerStarted","Data":"671456bc7912846bdfdf221593a6bd08d59c3f1933a3cbe4698a29745fb889d5"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.131896 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" event={"ID":"57dea3fa-cff1-4b2b-9193-2eef6ac70d32","Type":"ContainerStarted","Data":"f09722851a482f30a371a99e2f9d0e4614868958afea5e099e315357e5086e02"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.147766 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" event={"ID":"bb2dc677-b493-45e9-a3a0-fc805b303a11","Type":"ContainerStarted","Data":"0d01f0847c27c17794f2d946949a415f05168e81402eeef5c6072e9a3a353090"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.168495 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" event={"ID":"9a8a1bd5-206a-4154-886b-21f1b00906a6","Type":"ContainerStarted","Data":"dfd73573aa91b182c3740a6615ca12e2cc1261052759e8d9673d42b365fbe0e9"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.170489 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8zrgd" event={"ID":"c70614b5-a089-4e6e-8c59-06a01aaa2c54","Type":"ContainerStarted","Data":"d220570c8ef06e237155f08eead0368570879b9f2dd29a3063b02ed3a811c3fc"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.178956 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" event={"ID":"ab52f1c4-7d72-4d74-a16f-d49ed094b8e8","Type":"ContainerStarted","Data":"f86c6a07a77f9433bcfc795d20db72706883f14e1b6202be36f1a5fa29865ec2"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.195152 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.195996 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:05.69598206 +0000 UTC m=+155.680030428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.196106 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-6j5pp" podStartSLOduration=134.196092083 podStartE2EDuration="2m14.196092083s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:05.194714894 +0000 UTC m=+155.178763262" watchObservedRunningTime="2025-10-14 15:17:05.196092083 +0000 UTC m=+155.180140451" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.199520 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" event={"ID":"39778b3c-6ffc-417e-b9ff-50526b1fd867","Type":"ContainerStarted","Data":"13aa3f089758f31966bbcac140cbc88ecac93df7ab1a26b7aa53954f69bf65a7"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.199549 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" event={"ID":"39778b3c-6ffc-417e-b9ff-50526b1fd867","Type":"ContainerStarted","Data":"05ea3a0c58e6b4bd5d6026304518e812cd6e22b9aeeb1d22f545dca49be33035"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.200160 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.204989 4945 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-mk6kt container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.205027 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" podUID="39778b3c-6ffc-417e-b9ff-50526b1fd867" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.219475 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" event={"ID":"afa06ee0-ace4-4549-8bf4-2d02930cf2f5","Type":"ContainerStarted","Data":"7adb722db7eda62fccb0a0a3f708fa4eec067bbb038c87c4edb20fa4bc2f8a99"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.240820 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" event={"ID":"d8cdf424-f208-4e85-a001-0ffc3a1b27c2","Type":"ContainerStarted","Data":"2410768f444d97984a0a9166c8363cedc14473d7c636b748500583d2b048a2d8"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.258085 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" podStartSLOduration=134.258069148 podStartE2EDuration="2m14.258069148s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:05.229911335 +0000 UTC m=+155.213959703" watchObservedRunningTime="2025-10-14 15:17:05.258069148 +0000 UTC m=+155.242117516" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.265194 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" event={"ID":"388c7278-fae0-4de0-b646-49e46f88c517","Type":"ContainerStarted","Data":"4c3d93f5cd146165092d4f581c3db87c7447e44553555dafa4825d345b2f83dc"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.276991 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" event={"ID":"fd845cc3-5d83-4f84-9757-338b2eb40662","Type":"ContainerStarted","Data":"152b546388b56219b528ad8b4a033926c60f1ea40cd039fc5487507c51b3fe24"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.294545 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-s5dk6" event={"ID":"0ee69313-9ffd-4f06-aa3a-58037bd058de","Type":"ContainerStarted","Data":"929974d43973f180046b66a40ceeca45ec9d0998180b010fd2af0a3213dc3c87"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.296338 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.297484 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:05.797472627 +0000 UTC m=+155.781520995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.298072 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc" event={"ID":"82aa3bda-5c5a-4d70-826e-e0a5d358e3f2","Type":"ContainerStarted","Data":"f920cdff063715e39d29b34ac51430131bde946c16c01459a7d138aafcb86136"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.298110 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc" event={"ID":"82aa3bda-5c5a-4d70-826e-e0a5d358e3f2","Type":"ContainerStarted","Data":"3b7d92e6e79b9a57bec2c57cf687274d15bf7a485eec65d9e62f39e6821ca786"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.303129 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2j7ss" event={"ID":"cd5588a7-14c8-4a04-bccc-a22bc3edefaa","Type":"ContainerStarted","Data":"a0357389c41ba3e2f2ed9722591e25e1f1358bd4a893e64bf66af7dfcf56091d"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.310554 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" event={"ID":"ade216cb-9ef7-4dac-b35e-61dd3547b3c3","Type":"ContainerStarted","Data":"823accf68998fb845ed51b1624e491593eb2c4dd22998d9f2023fd1ed4509a4e"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.328010 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" event={"ID":"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6","Type":"ContainerStarted","Data":"50a5243678a5a07d3f08965e7725e0c72e1788f8dfa2fc37580c9d3214c43989"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.328057 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" event={"ID":"49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6","Type":"ContainerStarted","Data":"65628e70e30efbac52c4ef5329c1a42186ffa77f1b45adbe63d25739696ac8e2"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.328825 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.337016 4945 generic.go:334] "Generic (PLEG): container finished" podID="b683cb73-6094-43f6-b9f3-9f89691405c1" containerID="32506412de7c07bdedf42ae4fc10a9c454d20f84b73488deb1c897b3a50c842e" exitCode=0 Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.337091 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" event={"ID":"b683cb73-6094-43f6-b9f3-9f89691405c1","Type":"ContainerDied","Data":"32506412de7c07bdedf42ae4fc10a9c454d20f84b73488deb1c897b3a50c842e"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.337809 4945 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jtl86 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.337839 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" podUID="49e1bd46-d0ac-4c5d-9f71-f8d0be7915d6" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.401238 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.401352 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:05.90132446 +0000 UTC m=+155.885372828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.401713 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.402820 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:05.902805582 +0000 UTC m=+155.886853950 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.415275 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ccxsw" event={"ID":"06ecdccd-2560-48ea-a7f3-87e3e75488a0","Type":"ContainerStarted","Data":"354b14026fcbff4bf145f5c84b4f72443811c15403dc6c5671c2b01813577872"} Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.416710 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cv87 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.416764 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cv87" podUID="58804eb4-1016-427d-bcf6-fcfa9e8864df" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.423121 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.438292 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-hvp57" Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.502444 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.504115 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.004097483 +0000 UTC m=+155.988145861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.604599 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.605018 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.104990943 +0000 UTC m=+156.089039311 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.706647 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.706783 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.206762978 +0000 UTC m=+156.190811346 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.706912 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.707555 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.207525539 +0000 UTC m=+156.191573907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.807612 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.807798 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.30777139 +0000 UTC m=+156.291819758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.807961 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.808219 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.308206262 +0000 UTC m=+156.292254630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.909228 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.909433 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.409407751 +0000 UTC m=+156.393456119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.909773 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:05 crc kubenswrapper[4945]: E1014 15:17:05.916027 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.416014667 +0000 UTC m=+156.400063025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:05 crc kubenswrapper[4945]: I1014 15:17:05.974954 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-ccxsw" podStartSLOduration=134.974937536 podStartE2EDuration="2m14.974937536s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:05.973065023 +0000 UTC m=+155.957113391" watchObservedRunningTime="2025-10-14 15:17:05.974937536 +0000 UTC m=+155.958985904" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.010945 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:06 crc kubenswrapper[4945]: E1014 15:17:06.011230 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.511213327 +0000 UTC m=+156.495261695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.020972 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" podStartSLOduration=136.020942501 podStartE2EDuration="2m16.020942501s" podCreationTimestamp="2025-10-14 15:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.017560676 +0000 UTC m=+156.001609044" watchObservedRunningTime="2025-10-14 15:17:06.020942501 +0000 UTC m=+156.004990869" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.077915 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:06 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:06 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:06 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.077986 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.082812 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-2j7ss" podStartSLOduration=6.082791412 podStartE2EDuration="6.082791412s" podCreationTimestamp="2025-10-14 15:17:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.050817312 +0000 UTC m=+156.034865680" watchObservedRunningTime="2025-10-14 15:17:06.082791412 +0000 UTC m=+156.066839780" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.114467 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.118419 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8q5kj" podStartSLOduration=135.118394924 podStartE2EDuration="2m15.118394924s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.107303312 +0000 UTC m=+156.091351680" watchObservedRunningTime="2025-10-14 15:17:06.118394924 +0000 UTC m=+156.102443292" Oct 14 15:17:06 crc kubenswrapper[4945]: E1014 15:17:06.128712 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.628693304 +0000 UTC m=+156.612741672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.157563 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" podStartSLOduration=135.157549656 podStartE2EDuration="2m15.157549656s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.154947663 +0000 UTC m=+156.138996031" watchObservedRunningTime="2025-10-14 15:17:06.157549656 +0000 UTC m=+156.141598024" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.186701 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qnvrc" podStartSLOduration=135.186682926 podStartE2EDuration="2m15.186682926s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.184464524 +0000 UTC m=+156.168512892" watchObservedRunningTime="2025-10-14 15:17:06.186682926 +0000 UTC m=+156.170731294" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.217479 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:06 crc kubenswrapper[4945]: E1014 15:17:06.217910 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.717889975 +0000 UTC m=+156.701938343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.299224 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" podStartSLOduration=135.299206014 podStartE2EDuration="2m15.299206014s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.291190288 +0000 UTC m=+156.275238656" watchObservedRunningTime="2025-10-14 15:17:06.299206014 +0000 UTC m=+156.283254382" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.323967 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:06 crc kubenswrapper[4945]: E1014 15:17:06.324375 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.824362592 +0000 UTC m=+156.808410960 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.369997 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-frj5t" podStartSLOduration=136.369983296 podStartE2EDuration="2m16.369983296s" podCreationTimestamp="2025-10-14 15:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.368985068 +0000 UTC m=+156.353033436" watchObservedRunningTime="2025-10-14 15:17:06.369983296 +0000 UTC m=+156.354031664" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.399174 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" podStartSLOduration=135.399158097 podStartE2EDuration="2m15.399158097s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.38825473 +0000 UTC m=+156.372303098" watchObservedRunningTime="2025-10-14 15:17:06.399158097 +0000 UTC m=+156.383206455" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.425689 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:06 crc kubenswrapper[4945]: E1014 15:17:06.426021 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:06.926003013 +0000 UTC m=+156.910051381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.503624 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" event={"ID":"748d0e0e-3657-4ad3-8918-c0d0773e06e6","Type":"ContainerStarted","Data":"f7d08b723de14084a71a3f6c2d58a70f601e9c272f29e43629121ca81609d9e7"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.505138 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.527246 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:06 crc kubenswrapper[4945]: E1014 15:17:06.527569 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:07.027544441 +0000 UTC m=+157.011592809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.535992 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-8zrgd" podStartSLOduration=7.535974609 podStartE2EDuration="7.535974609s" podCreationTimestamp="2025-10-14 15:16:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.428693089 +0000 UTC m=+156.412741477" watchObservedRunningTime="2025-10-14 15:17:06.535974609 +0000 UTC m=+156.520022977" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.537416 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" podStartSLOduration=135.537410899 podStartE2EDuration="2m15.537410899s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.53530875 +0000 UTC m=+156.519357118" watchObservedRunningTime="2025-10-14 15:17:06.537410899 +0000 UTC m=+156.521459267" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.545691 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-s5dk6" event={"ID":"0ee69313-9ffd-4f06-aa3a-58037bd058de","Type":"ContainerStarted","Data":"e03e38082410bd0d85344ed273925715b32d4a076ed36cd17d8a8a5777b90aaa"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.545726 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-s5dk6" event={"ID":"0ee69313-9ffd-4f06-aa3a-58037bd058de","Type":"ContainerStarted","Data":"bf29a89734010f97a0ee7f30836fc02b81884eb8b7e86ec4439c74cd16d6f5c0"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.571108 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r8hlp" event={"ID":"57dea3fa-cff1-4b2b-9193-2eef6ac70d32","Type":"ContainerStarted","Data":"0d37e25e268cec6168b3e4b3f5587fda813c548770839fe58465bfe049d6756a"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.581175 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" event={"ID":"ab52f1c4-7d72-4d74-a16f-d49ed094b8e8","Type":"ContainerStarted","Data":"14550354e62ef7457ec18dd1b6d17a0696ee41a5255bbd1a7a8d3c661f370ed7"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.582125 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-s5dk6" podStartSLOduration=135.582111387 podStartE2EDuration="2m15.582111387s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.579505034 +0000 UTC m=+156.563553402" watchObservedRunningTime="2025-10-14 15:17:06.582111387 +0000 UTC m=+156.566159745" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.606477 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" event={"ID":"fd845cc3-5d83-4f84-9757-338b2eb40662","Type":"ContainerStarted","Data":"4ed392261ab822faa75c026f68760d35c5a59f750e68e7448105ba44c20147e2"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.606532 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" event={"ID":"fd845cc3-5d83-4f84-9757-338b2eb40662","Type":"ContainerStarted","Data":"3bd3d175af05b8406f0a66ee607be884db20ffa93f8ce4cf77f7731c6b2937dd"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.616259 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5v9kf" podStartSLOduration=135.616244548 podStartE2EDuration="2m15.616244548s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.60602536 +0000 UTC m=+156.590073728" watchObservedRunningTime="2025-10-14 15:17:06.616244548 +0000 UTC m=+156.600292916" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.628397 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:06 crc kubenswrapper[4945]: E1014 15:17:06.629426 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:07.129395268 +0000 UTC m=+157.113443646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.635572 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" event={"ID":"ade216cb-9ef7-4dac-b35e-61dd3547b3c3","Type":"ContainerStarted","Data":"b58c89b252ded297145b99d7d80289fcc595958257d3ccc99bf9a955dbce6432"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.635612 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" event={"ID":"ade216cb-9ef7-4dac-b35e-61dd3547b3c3","Type":"ContainerStarted","Data":"4fc04b9f21567cf31ad7abeebb4bebb1a35673ce3d8a68563329172b7a9a28d9"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.642968 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dmdlz" podStartSLOduration=135.64295582 podStartE2EDuration="2m15.64295582s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.642693432 +0000 UTC m=+156.626741800" watchObservedRunningTime="2025-10-14 15:17:06.64295582 +0000 UTC m=+156.627004188" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.652226 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" event={"ID":"98994fb5-4f28-4381-9ae2-4c734bf79c5c","Type":"ContainerStarted","Data":"325e630298c84c6bd6ad7affff339d43dc7e628aa02b16abf28c73bdbc17bb9d"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.653356 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.656498 4945 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2kwtj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.656567 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" podUID="98994fb5-4f28-4381-9ae2-4c734bf79c5c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.681092 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" event={"ID":"877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2","Type":"ContainerStarted","Data":"7025dfd3fea3bb994118a489765880fe0cda38fc6eec0b38868005b7d5ec2be7"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.681135 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" event={"ID":"877b02e6-9caf-4066-bc2c-4d4d3cf8b7e2","Type":"ContainerStarted","Data":"efe60b49c6edb559c8f1f34f7cacecbf8c1b88f2e6468f1416b391957ad17f53"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.689228 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" event={"ID":"a98ffecf-8060-4d01-8bfe-e1f7c8f04b46","Type":"ContainerStarted","Data":"2dba665be4194fdb99171e2f21bdfca9c8ab00448b549c31d1615aa2ff8f0eee"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.699401 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-m8jbn" podStartSLOduration=135.699383558 podStartE2EDuration="2m15.699383558s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.698468753 +0000 UTC m=+156.682517121" watchObservedRunningTime="2025-10-14 15:17:06.699383558 +0000 UTC m=+156.683431926" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.713858 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9" event={"ID":"b47bfeba-2bc6-4c04-bd5a-a02b10ef8478","Type":"ContainerStarted","Data":"202c5077cdaa1313b5c9fe011836b6224bfb4e35ddee5ad169dc5734c5707cd9"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.730253 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:06 crc kubenswrapper[4945]: E1014 15:17:06.732204 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:07.232187292 +0000 UTC m=+157.216235650 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.733267 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-9chhc" podStartSLOduration=135.733251122 podStartE2EDuration="2m15.733251122s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.732769488 +0000 UTC m=+156.716817856" watchObservedRunningTime="2025-10-14 15:17:06.733251122 +0000 UTC m=+156.717299490" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.734778 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-45pg9" event={"ID":"d5913ac3-e4f4-4397-bc8c-3fd7ab4fee4d","Type":"ContainerStarted","Data":"044c0d085a84ec420278c8e8ed5fd9fdba6e528d8cf36f9f5456831b02eae8c1"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.734921 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-45pg9" event={"ID":"d5913ac3-e4f4-4397-bc8c-3fd7ab4fee4d","Type":"ContainerStarted","Data":"d309db484b0db5f32559b296aa76f18a85f3ee1e9c102bffb5eaf05f46340b92"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.744487 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2j7ss" event={"ID":"cd5588a7-14c8-4a04-bccc-a22bc3edefaa","Type":"ContainerStarted","Data":"b110e031fc247c344bc94f3ab3bd6067dd71b4d17c5f17709d8bbf4d570bf179"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.756376 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-h98fq" event={"ID":"c788e539-e984-4afb-ac5f-d216466ff704","Type":"ContainerStarted","Data":"b708993b6e859c416116039cfc86ed3a85d18ec0d9ca5524e7bf46e50c24c840"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.756430 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-h98fq" event={"ID":"c788e539-e984-4afb-ac5f-d216466ff704","Type":"ContainerStarted","Data":"1d07705ae2f400677bda2e7b1464b84f09053d94f3974d6aa89655cfd301a21c"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.758609 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c25w8" podStartSLOduration=135.75771236 podStartE2EDuration="2m15.75771236s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.757324159 +0000 UTC m=+156.741372527" watchObservedRunningTime="2025-10-14 15:17:06.75771236 +0000 UTC m=+156.741760728" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.790403 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" event={"ID":"9a8a1bd5-206a-4154-886b-21f1b00906a6","Type":"ContainerStarted","Data":"fa3be5c0edbc5304e6ed88f36c687757106d962d29561b728e425d9f6df438d4"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.800612 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" event={"ID":"3ec13203-2a26-4de1-b02f-30c58d8ab5a1","Type":"ContainerStarted","Data":"fed595aecd66fbfee88bb58793a77468af4415631ac0388fffa2a9f78cd7cf14"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.801810 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" podStartSLOduration=135.801797961 podStartE2EDuration="2m15.801797961s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.801320158 +0000 UTC m=+156.785368526" watchObservedRunningTime="2025-10-14 15:17:06.801797961 +0000 UTC m=+156.785846319" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.823337 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" podStartSLOduration=126.823320987 podStartE2EDuration="2m6.823320987s" podCreationTimestamp="2025-10-14 15:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.821747993 +0000 UTC m=+156.805796361" watchObservedRunningTime="2025-10-14 15:17:06.823320987 +0000 UTC m=+156.807369375" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.829320 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ccxsw" event={"ID":"06ecdccd-2560-48ea-a7f3-87e3e75488a0","Type":"ContainerStarted","Data":"b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.831904 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:06 crc kubenswrapper[4945]: E1014 15:17:06.833351 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:07.333332239 +0000 UTC m=+157.317380607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.854848 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-h98fq" podStartSLOduration=135.854829774 podStartE2EDuration="2m15.854829774s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.843887656 +0000 UTC m=+156.827936024" watchObservedRunningTime="2025-10-14 15:17:06.854829774 +0000 UTC m=+156.838878142" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.885857 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfgxp" podStartSLOduration=135.885840637 podStartE2EDuration="2m15.885840637s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.882438681 +0000 UTC m=+156.866487049" watchObservedRunningTime="2025-10-14 15:17:06.885840637 +0000 UTC m=+156.869889065" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.907154 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" event={"ID":"81c3c0f8-ad71-4a4b-8193-95051f481416","Type":"ContainerStarted","Data":"8e8c36ca97a9e95abc11e3ca6e204db1978b51b5603560cc778a3c476545f5c8"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.935975 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.936963 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" event={"ID":"bb2dc677-b493-45e9-a3a0-fc805b303a11","Type":"ContainerStarted","Data":"ae8a2fa94cc28d3293d542b9c341d66d82a8ae27524a48b8a5d33cc2981e70de"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.937004 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" event={"ID":"bb2dc677-b493-45e9-a3a0-fc805b303a11","Type":"ContainerStarted","Data":"f869777dfebf1730156395f014d54f1195236bbdb72e3cc6467ab58601e70c4c"} Oct 14 15:17:06 crc kubenswrapper[4945]: E1014 15:17:06.937605 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:07.437590024 +0000 UTC m=+157.421638392 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.949502 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-45pg9" podStartSLOduration=135.949485328 podStartE2EDuration="2m15.949485328s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.920165573 +0000 UTC m=+156.904213931" watchObservedRunningTime="2025-10-14 15:17:06.949485328 +0000 UTC m=+156.933533696" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.950097 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" event={"ID":"b683cb73-6094-43f6-b9f3-9f89691405c1","Type":"ContainerStarted","Data":"75a32288a18f864f23d68675a073d2441b42f39fa92bc1ec1ee5a330cb7e723a"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.978783 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" event={"ID":"fdb8607a-0a03-4e36-bf27-080330220d71","Type":"ContainerStarted","Data":"010f32881e94fa8c751bf65929efd118986c1484059b6dabb3ea9f76b26f58be"} Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.979618 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.987603 4945 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-9qkjs container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.987770 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" podUID="fdb8607a-0a03-4e36-bf27-080330220d71" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.994997 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jphx9" podStartSLOduration=135.994981239 podStartE2EDuration="2m15.994981239s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.950099536 +0000 UTC m=+156.934147904" watchObservedRunningTime="2025-10-14 15:17:06.994981239 +0000 UTC m=+156.979029607" Oct 14 15:17:06 crc kubenswrapper[4945]: I1014 15:17:06.996165 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" event={"ID":"afa06ee0-ace4-4549-8bf4-2d02930cf2f5","Type":"ContainerStarted","Data":"94dc7e2ad2b6734f3c2d3b1830ce6f410cc4163b085be70c3a3bfa022f376c11"} Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:06.997325 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jpztt" podStartSLOduration=135.997314165 podStartE2EDuration="2m15.997314165s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:06.995750881 +0000 UTC m=+156.979799249" watchObservedRunningTime="2025-10-14 15:17:06.997314165 +0000 UTC m=+156.981362533" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.010895 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jtl86" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.022612 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mk6kt" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.038042 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:07 crc kubenswrapper[4945]: E1014 15:17:07.039105 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:07.539088971 +0000 UTC m=+157.523137339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.048315 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" podStartSLOduration=137.04828987 podStartE2EDuration="2m17.04828987s" podCreationTimestamp="2025-10-14 15:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:07.046315784 +0000 UTC m=+157.030364152" watchObservedRunningTime="2025-10-14 15:17:07.04828987 +0000 UTC m=+157.032338238" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.048563 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:07 crc kubenswrapper[4945]: E1014 15:17:07.055856 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:07.555842152 +0000 UTC m=+157.539890510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.070527 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:07 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:07 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:07 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.070860 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.149980 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:07 crc kubenswrapper[4945]: E1014 15:17:07.150253 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:07.650239029 +0000 UTC m=+157.634287397 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.219781 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" podStartSLOduration=136.219761796 podStartE2EDuration="2m16.219761796s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:07.217721719 +0000 UTC m=+157.201770087" watchObservedRunningTime="2025-10-14 15:17:07.219761796 +0000 UTC m=+157.203810164" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.251912 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.252459 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" podStartSLOduration=136.252442976 podStartE2EDuration="2m16.252442976s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:07.251436468 +0000 UTC m=+157.235484836" watchObservedRunningTime="2025-10-14 15:17:07.252442976 +0000 UTC m=+157.236491344" Oct 14 15:17:07 crc kubenswrapper[4945]: E1014 15:17:07.252644 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:07.752631812 +0000 UTC m=+157.736680180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.288943 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nb6gw" podStartSLOduration=136.288927523 podStartE2EDuration="2m16.288927523s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:07.286424363 +0000 UTC m=+157.270472731" watchObservedRunningTime="2025-10-14 15:17:07.288927523 +0000 UTC m=+157.272975891" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.353313 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:07 crc kubenswrapper[4945]: E1014 15:17:07.353664 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:07.853646805 +0000 UTC m=+157.837695173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.454201 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:07 crc kubenswrapper[4945]: E1014 15:17:07.454644 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:07.954626508 +0000 UTC m=+157.938674876 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.476474 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.551045 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.551342 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.552726 4945 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-mnqtf container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.11:8443/livez\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.552764 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" podUID="b683cb73-6094-43f6-b9f3-9f89691405c1" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.11:8443/livez\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.555540 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:07 crc kubenswrapper[4945]: E1014 15:17:07.555819 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.055806486 +0000 UTC m=+158.039854854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.619709 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.619994 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.621225 4945 patch_prober.go:28] interesting pod/apiserver-76f77b778f-8f2dm container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.14:8443/livez\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.621260 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" podUID="81c3c0f8-ad71-4a4b-8193-95051f481416" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.14:8443/livez\": dial tcp 10.217.0.14:8443: connect: connection refused" Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.657043 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:07 crc kubenswrapper[4945]: E1014 15:17:07.657304 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.157294292 +0000 UTC m=+158.141342660 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.757710 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:07 crc kubenswrapper[4945]: E1014 15:17:07.758004 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.257989867 +0000 UTC m=+158.242038235 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.859133 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:07 crc kubenswrapper[4945]: E1014 15:17:07.859515 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.359499934 +0000 UTC m=+158.343548302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.960283 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:07 crc kubenswrapper[4945]: E1014 15:17:07.960435 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.460409845 +0000 UTC m=+158.444458213 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:07 crc kubenswrapper[4945]: I1014 15:17:07.960499 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:07 crc kubenswrapper[4945]: E1014 15:17:07.960842 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.460820296 +0000 UTC m=+158.444868664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.009921 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" event={"ID":"81c3c0f8-ad71-4a4b-8193-95051f481416","Type":"ContainerStarted","Data":"77b277f04048253ce03cccc17d214fbe5e7adb793a767e2d13f400f1552f252c"} Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.018811 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" event={"ID":"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc","Type":"ContainerStarted","Data":"e8e7de3d6631228370cc8cd33c8fee3c69752f56efb313c990b9a708266bcc62"} Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.018857 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" event={"ID":"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc","Type":"ContainerStarted","Data":"f1f239b3d054d621508af8649e50ec495839fc5e2cf1564bc9e2582b36e0b1dd"} Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.019616 4945 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2kwtj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.019665 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" podUID="98994fb5-4f28-4381-9ae2-4c734bf79c5c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.036868 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9qkjs" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.061142 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.061576 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.561559512 +0000 UTC m=+158.545607880 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.064208 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.069387 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.569337921 +0000 UTC m=+158.553386279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.078714 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:08 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:08 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:08 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.086081 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.161684 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vs4wr"] Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.162644 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.164956 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.165094 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.165278 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.665254241 +0000 UTC m=+158.649302609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.165894 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.168889 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.668854792 +0000 UTC m=+158.652903260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.179992 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vs4wr"] Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.267338 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.267541 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.767483219 +0000 UTC m=+158.751531587 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.268125 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.268162 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkwqk\" (UniqueName: \"kubernetes.io/projected/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-kube-api-access-lkwqk\") pod \"community-operators-vs4wr\" (UID: \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\") " pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.268229 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-utilities\") pod \"community-operators-vs4wr\" (UID: \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\") " pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.268547 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-catalog-content\") pod \"community-operators-vs4wr\" (UID: \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\") " pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.268720 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.768702823 +0000 UTC m=+158.752751191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.307394 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bvplf" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.369336 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.369455 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.869433968 +0000 UTC m=+158.853482336 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.369564 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-catalog-content\") pod \"community-operators-vs4wr\" (UID: \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\") " pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.369674 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.369714 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkwqk\" (UniqueName: \"kubernetes.io/projected/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-kube-api-access-lkwqk\") pod \"community-operators-vs4wr\" (UID: \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\") " pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.369757 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-utilities\") pod \"community-operators-vs4wr\" (UID: \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\") " pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.370105 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.870089257 +0000 UTC m=+158.854137625 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.370197 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-utilities\") pod \"community-operators-vs4wr\" (UID: \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\") " pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.371191 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-catalog-content\") pod \"community-operators-vs4wr\" (UID: \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\") " pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.402223 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tfj4f"] Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.403716 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.411076 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.411944 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkwqk\" (UniqueName: \"kubernetes.io/projected/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-kube-api-access-lkwqk\") pod \"community-operators-vs4wr\" (UID: \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\") " pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.414343 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tfj4f"] Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.472324 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.472528 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec783233-16ca-4e17-8923-71455b2056d7-catalog-content\") pod \"certified-operators-tfj4f\" (UID: \"ec783233-16ca-4e17-8923-71455b2056d7\") " pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.472595 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klwwd\" (UniqueName: \"kubernetes.io/projected/ec783233-16ca-4e17-8923-71455b2056d7-kube-api-access-klwwd\") pod \"certified-operators-tfj4f\" (UID: \"ec783233-16ca-4e17-8923-71455b2056d7\") " pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.472617 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec783233-16ca-4e17-8923-71455b2056d7-utilities\") pod \"certified-operators-tfj4f\" (UID: \"ec783233-16ca-4e17-8923-71455b2056d7\") " pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.472734 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:08.972720376 +0000 UTC m=+158.956768744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.485023 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.509179 4945 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.550742 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wvfqb"] Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.551635 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.573706 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klwwd\" (UniqueName: \"kubernetes.io/projected/ec783233-16ca-4e17-8923-71455b2056d7-kube-api-access-klwwd\") pod \"certified-operators-tfj4f\" (UID: \"ec783233-16ca-4e17-8923-71455b2056d7\") " pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.574019 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec783233-16ca-4e17-8923-71455b2056d7-utilities\") pod \"certified-operators-tfj4f\" (UID: \"ec783233-16ca-4e17-8923-71455b2056d7\") " pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.574056 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.574102 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec783233-16ca-4e17-8923-71455b2056d7-catalog-content\") pod \"certified-operators-tfj4f\" (UID: \"ec783233-16ca-4e17-8923-71455b2056d7\") " pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.574478 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec783233-16ca-4e17-8923-71455b2056d7-catalog-content\") pod \"certified-operators-tfj4f\" (UID: \"ec783233-16ca-4e17-8923-71455b2056d7\") " pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.574650 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec783233-16ca-4e17-8923-71455b2056d7-utilities\") pod \"certified-operators-tfj4f\" (UID: \"ec783233-16ca-4e17-8923-71455b2056d7\") " pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.573933 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wvfqb"] Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.574721 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:09.074709437 +0000 UTC m=+159.058757805 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.614283 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klwwd\" (UniqueName: \"kubernetes.io/projected/ec783233-16ca-4e17-8923-71455b2056d7-kube-api-access-klwwd\") pod \"certified-operators-tfj4f\" (UID: \"ec783233-16ca-4e17-8923-71455b2056d7\") " pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.675262 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.675435 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:09.175417672 +0000 UTC m=+159.159466040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.675482 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2424e499-800f-481e-95e2-84257025d221-catalog-content\") pod \"community-operators-wvfqb\" (UID: \"2424e499-800f-481e-95e2-84257025d221\") " pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.675525 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2424e499-800f-481e-95e2-84257025d221-utilities\") pod \"community-operators-wvfqb\" (UID: \"2424e499-800f-481e-95e2-84257025d221\") " pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.675546 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.675585 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm4j2\" (UniqueName: \"kubernetes.io/projected/2424e499-800f-481e-95e2-84257025d221-kube-api-access-jm4j2\") pod \"community-operators-wvfqb\" (UID: \"2424e499-800f-481e-95e2-84257025d221\") " pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.675855 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:09.175848404 +0000 UTC m=+159.159896762 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.736647 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.748644 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8z5p9"] Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.749720 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.756128 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8z5p9"] Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.777537 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.777821 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm4j2\" (UniqueName: \"kubernetes.io/projected/2424e499-800f-481e-95e2-84257025d221-kube-api-access-jm4j2\") pod \"community-operators-wvfqb\" (UID: \"2424e499-800f-481e-95e2-84257025d221\") " pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.777898 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2424e499-800f-481e-95e2-84257025d221-catalog-content\") pod \"community-operators-wvfqb\" (UID: \"2424e499-800f-481e-95e2-84257025d221\") " pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.777937 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2424e499-800f-481e-95e2-84257025d221-utilities\") pod \"community-operators-wvfqb\" (UID: \"2424e499-800f-481e-95e2-84257025d221\") " pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.778682 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:09.278651237 +0000 UTC m=+159.262699655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.779082 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2424e499-800f-481e-95e2-84257025d221-utilities\") pod \"community-operators-wvfqb\" (UID: \"2424e499-800f-481e-95e2-84257025d221\") " pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.779222 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2424e499-800f-481e-95e2-84257025d221-catalog-content\") pod \"community-operators-wvfqb\" (UID: \"2424e499-800f-481e-95e2-84257025d221\") " pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.798406 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm4j2\" (UniqueName: \"kubernetes.io/projected/2424e499-800f-481e-95e2-84257025d221-kube-api-access-jm4j2\") pod \"community-operators-wvfqb\" (UID: \"2424e499-800f-481e-95e2-84257025d221\") " pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.842294 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vs4wr"] Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.866230 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.879551 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-utilities\") pod \"certified-operators-8z5p9\" (UID: \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\") " pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.879621 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.879651 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-catalog-content\") pod \"certified-operators-8z5p9\" (UID: \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\") " pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.880023 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 15:17:09.380008331 +0000 UTC m=+159.364056699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4bf5h" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.881313 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmnsl\" (UniqueName: \"kubernetes.io/projected/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-kube-api-access-tmnsl\") pod \"certified-operators-8z5p9\" (UID: \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\") " pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.966740 4945 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-14T15:17:08.509202813Z","Handler":null,"Name":""} Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.981988 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.982589 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-utilities\") pod \"certified-operators-8z5p9\" (UID: \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\") " pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.982678 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-catalog-content\") pod \"certified-operators-8z5p9\" (UID: \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\") " pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.982721 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmnsl\" (UniqueName: \"kubernetes.io/projected/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-kube-api-access-tmnsl\") pod \"certified-operators-8z5p9\" (UID: \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\") " pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:17:08 crc kubenswrapper[4945]: E1014 15:17:08.983404 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 15:17:09.483389521 +0000 UTC m=+159.467437889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.983765 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-utilities\") pod \"certified-operators-8z5p9\" (UID: \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\") " pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.983996 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-catalog-content\") pod \"certified-operators-8z5p9\" (UID: \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\") " pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.986217 4945 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 14 15:17:08 crc kubenswrapper[4945]: I1014 15:17:08.986247 4945 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.043949 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tfj4f"] Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.048611 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmnsl\" (UniqueName: \"kubernetes.io/projected/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-kube-api-access-tmnsl\") pod \"certified-operators-8z5p9\" (UID: \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\") " pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.056020 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" event={"ID":"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc","Type":"ContainerStarted","Data":"536ae6f3b122cc265e68a659be546a536ea5395841f7364ecdd71ca87bd675a9"} Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.069581 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:09 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:09 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:09 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.069666 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.070135 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vs4wr" event={"ID":"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039","Type":"ContainerStarted","Data":"0a555505cbef0d00e1a2d3f9639cb30f5473cd21b8b987669ddf5f278d3a5c50"} Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.084369 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.089200 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.093077 4945 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.093122 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.106752 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.151557 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4bf5h\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.186584 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.217797 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.253202 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.353169 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wvfqb"] Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.527586 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4bf5h"] Oct 14 15:17:09 crc kubenswrapper[4945]: W1014 15:17:09.536361 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09ab4e69_a6c2_41f6_a1e6_d59919746fd9.slice/crio-48d54052f7678f8046e4f38ae0044d168dc1891b5d867e7c2e8d4ebc001bcfc2 WatchSource:0}: Error finding container 48d54052f7678f8046e4f38ae0044d168dc1891b5d867e7c2e8d4ebc001bcfc2: Status 404 returned error can't find the container with id 48d54052f7678f8046e4f38ae0044d168dc1891b5d867e7c2e8d4ebc001bcfc2 Oct 14 15:17:09 crc kubenswrapper[4945]: I1014 15:17:09.640258 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8z5p9"] Oct 14 15:17:09 crc kubenswrapper[4945]: W1014 15:17:09.646400 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba2d8579_0ecb_46a2_950b_c48d660ed7ab.slice/crio-d11fd99ea853960d888477eda5794744f2c555fadc4a700a60bf908adf98ec0d WatchSource:0}: Error finding container d11fd99ea853960d888477eda5794744f2c555fadc4a700a60bf908adf98ec0d: Status 404 returned error can't find the container with id d11fd99ea853960d888477eda5794744f2c555fadc4a700a60bf908adf98ec0d Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.063176 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:10 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:10 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:10 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.063494 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.075367 4945 generic.go:334] "Generic (PLEG): container finished" podID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" containerID="1eb0e4e31d78e600e87e9d170c0db58f48e5f04c1291d986b32e5ea2ad0d324b" exitCode=0 Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.075415 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z5p9" event={"ID":"ba2d8579-0ecb-46a2-950b-c48d660ed7ab","Type":"ContainerDied","Data":"1eb0e4e31d78e600e87e9d170c0db58f48e5f04c1291d986b32e5ea2ad0d324b"} Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.075463 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z5p9" event={"ID":"ba2d8579-0ecb-46a2-950b-c48d660ed7ab","Type":"ContainerStarted","Data":"d11fd99ea853960d888477eda5794744f2c555fadc4a700a60bf908adf98ec0d"} Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.076815 4945 generic.go:334] "Generic (PLEG): container finished" podID="ec783233-16ca-4e17-8923-71455b2056d7" containerID="5ec5f64ebde6d98946d7c2a23464c2e4bc65b08e560315ac1ba8cd98f71a2600" exitCode=0 Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.076972 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfj4f" event={"ID":"ec783233-16ca-4e17-8923-71455b2056d7","Type":"ContainerDied","Data":"5ec5f64ebde6d98946d7c2a23464c2e4bc65b08e560315ac1ba8cd98f71a2600"} Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.077018 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfj4f" event={"ID":"ec783233-16ca-4e17-8923-71455b2056d7","Type":"ContainerStarted","Data":"0232d42e0d70e53fbbc53df8ffceb40119496e95f239e46502987f7c682c35eb"} Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.077132 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.079729 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" event={"ID":"c04b76f5-5b1c-40fc-9bc2-b0fc3b824bfc","Type":"ContainerStarted","Data":"00e9662959b9fe3be056e580c1e53e4b4bfa0aa6edfc266dd544f0eb27f027a9"} Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.082476 4945 generic.go:334] "Generic (PLEG): container finished" podID="3ec13203-2a26-4de1-b02f-30c58d8ab5a1" containerID="fed595aecd66fbfee88bb58793a77468af4415631ac0388fffa2a9f78cd7cf14" exitCode=0 Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.082533 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" event={"ID":"3ec13203-2a26-4de1-b02f-30c58d8ab5a1","Type":"ContainerDied","Data":"fed595aecd66fbfee88bb58793a77468af4415631ac0388fffa2a9f78cd7cf14"} Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.083838 4945 generic.go:334] "Generic (PLEG): container finished" podID="2424e499-800f-481e-95e2-84257025d221" containerID="2c7420b342a6bc99d2d76dd72f7ba10db3bf359f30009ab6835a99f8b3a82843" exitCode=0 Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.083906 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvfqb" event={"ID":"2424e499-800f-481e-95e2-84257025d221","Type":"ContainerDied","Data":"2c7420b342a6bc99d2d76dd72f7ba10db3bf359f30009ab6835a99f8b3a82843"} Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.083929 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvfqb" event={"ID":"2424e499-800f-481e-95e2-84257025d221","Type":"ContainerStarted","Data":"2d43822aea2c9322a313dc58d7efd58ca721cb31d3b22fc37e09a961af79c440"} Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.086281 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" event={"ID":"09ab4e69-a6c2-41f6-a1e6-d59919746fd9","Type":"ContainerStarted","Data":"24e5d099d02a577a3bee77f291fa8fc870b73ecc8d22747d76e85c70d8a8482c"} Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.086341 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" event={"ID":"09ab4e69-a6c2-41f6-a1e6-d59919746fd9","Type":"ContainerStarted","Data":"48d54052f7678f8046e4f38ae0044d168dc1891b5d867e7c2e8d4ebc001bcfc2"} Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.086364 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.087668 4945 generic.go:334] "Generic (PLEG): container finished" podID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" containerID="a6150ee7d76eb524532781a91fe5de51725e6b65e0319f922a369bb556d79c74" exitCode=0 Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.087723 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vs4wr" event={"ID":"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039","Type":"ContainerDied","Data":"a6150ee7d76eb524532781a91fe5de51725e6b65e0319f922a369bb556d79c74"} Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.113716 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" podStartSLOduration=139.113698147 podStartE2EDuration="2m19.113698147s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:10.113408958 +0000 UTC m=+160.097457326" watchObservedRunningTime="2025-10-14 15:17:10.113698147 +0000 UTC m=+160.097746515" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.187768 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.188542 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.190057 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.191655 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.199244 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.214530 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-tgs6r" podStartSLOduration=10.214505534 podStartE2EDuration="10.214505534s" podCreationTimestamp="2025-10-14 15:17:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:10.209779351 +0000 UTC m=+160.193827739" watchObservedRunningTime="2025-10-14 15:17:10.214505534 +0000 UTC m=+160.198553902" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.301040 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/625cab2a-e7e1-44f7-b06a-61651f5c7d88-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"625cab2a-e7e1-44f7-b06a-61651f5c7d88\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.301152 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/625cab2a-e7e1-44f7-b06a-61651f5c7d88-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"625cab2a-e7e1-44f7-b06a-61651f5c7d88\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.352364 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b27lt"] Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.353556 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.356529 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.364723 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b27lt"] Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.402203 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/625cab2a-e7e1-44f7-b06a-61651f5c7d88-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"625cab2a-e7e1-44f7-b06a-61651f5c7d88\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.402250 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-utilities\") pod \"redhat-marketplace-b27lt\" (UID: \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\") " pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.402278 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/625cab2a-e7e1-44f7-b06a-61651f5c7d88-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"625cab2a-e7e1-44f7-b06a-61651f5c7d88\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.402300 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-catalog-content\") pod \"redhat-marketplace-b27lt\" (UID: \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\") " pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.402344 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mm4p\" (UniqueName: \"kubernetes.io/projected/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-kube-api-access-4mm4p\") pod \"redhat-marketplace-b27lt\" (UID: \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\") " pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.402624 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/625cab2a-e7e1-44f7-b06a-61651f5c7d88-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"625cab2a-e7e1-44f7-b06a-61651f5c7d88\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.424627 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/625cab2a-e7e1-44f7-b06a-61651f5c7d88-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"625cab2a-e7e1-44f7-b06a-61651f5c7d88\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.502382 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.503310 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-catalog-content\") pod \"redhat-marketplace-b27lt\" (UID: \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\") " pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.503361 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mm4p\" (UniqueName: \"kubernetes.io/projected/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-kube-api-access-4mm4p\") pod \"redhat-marketplace-b27lt\" (UID: \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\") " pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.503424 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-utilities\") pod \"redhat-marketplace-b27lt\" (UID: \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\") " pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.503810 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-utilities\") pod \"redhat-marketplace-b27lt\" (UID: \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\") " pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.504122 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-catalog-content\") pod \"redhat-marketplace-b27lt\" (UID: \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\") " pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.544912 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mm4p\" (UniqueName: \"kubernetes.io/projected/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-kube-api-access-4mm4p\") pod \"redhat-marketplace-b27lt\" (UID: \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\") " pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.666863 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.744633 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-prqws"] Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.745616 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.772760 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.774147 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-prqws"] Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.807046 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9gzn\" (UniqueName: \"kubernetes.io/projected/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-kube-api-access-j9gzn\") pod \"redhat-marketplace-prqws\" (UID: \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\") " pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.807106 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-catalog-content\") pod \"redhat-marketplace-prqws\" (UID: \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\") " pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.807345 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-utilities\") pod \"redhat-marketplace-prqws\" (UID: \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\") " pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.811055 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 14 15:17:10 crc kubenswrapper[4945]: W1014 15:17:10.814443 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod625cab2a_e7e1_44f7_b06a_61651f5c7d88.slice/crio-bc5c3d0f43b5ecb81b5dab38a1b01943a7a0a692e0e486d0d3cff7bd1edfc6cb WatchSource:0}: Error finding container bc5c3d0f43b5ecb81b5dab38a1b01943a7a0a692e0e486d0d3cff7bd1edfc6cb: Status 404 returned error can't find the container with id bc5c3d0f43b5ecb81b5dab38a1b01943a7a0a692e0e486d0d3cff7bd1edfc6cb Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.888254 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b27lt"] Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.910860 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-catalog-content\") pod \"redhat-marketplace-prqws\" (UID: \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\") " pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.910959 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-utilities\") pod \"redhat-marketplace-prqws\" (UID: \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\") " pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.910992 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9gzn\" (UniqueName: \"kubernetes.io/projected/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-kube-api-access-j9gzn\") pod \"redhat-marketplace-prqws\" (UID: \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\") " pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.911476 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-catalog-content\") pod \"redhat-marketplace-prqws\" (UID: \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\") " pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.915603 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-utilities\") pod \"redhat-marketplace-prqws\" (UID: \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\") " pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:17:10 crc kubenswrapper[4945]: I1014 15:17:10.933176 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9gzn\" (UniqueName: \"kubernetes.io/projected/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-kube-api-access-j9gzn\") pod \"redhat-marketplace-prqws\" (UID: \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\") " pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.061932 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.062723 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:11 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:11 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:11 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.062768 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.093334 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"625cab2a-e7e1-44f7-b06a-61651f5c7d88","Type":"ContainerStarted","Data":"bc5c3d0f43b5ecb81b5dab38a1b01943a7a0a692e0e486d0d3cff7bd1edfc6cb"} Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.094885 4945 generic.go:334] "Generic (PLEG): container finished" podID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" containerID="20c90023e137420bf01e365fc2c7cfb6d03619a7f5f3463dcae74761b09a6175" exitCode=0 Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.095020 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b27lt" event={"ID":"27eaa4e2-7ffd-41b3-9d3d-418b43669de8","Type":"ContainerDied","Data":"20c90023e137420bf01e365fc2c7cfb6d03619a7f5f3463dcae74761b09a6175"} Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.095082 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b27lt" event={"ID":"27eaa4e2-7ffd-41b3-9d3d-418b43669de8","Type":"ContainerStarted","Data":"6384aaee345c53151ec52a0cdcab05d97e9e340f03023a12f14c7effa334b6fa"} Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.325784 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.360970 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-prqws"] Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.379548 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5lphm"] Oct 14 15:17:11 crc kubenswrapper[4945]: E1014 15:17:11.379814 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ec13203-2a26-4de1-b02f-30c58d8ab5a1" containerName="collect-profiles" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.379831 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ec13203-2a26-4de1-b02f-30c58d8ab5a1" containerName="collect-profiles" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.379980 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ec13203-2a26-4de1-b02f-30c58d8ab5a1" containerName="collect-profiles" Oct 14 15:17:11 crc kubenswrapper[4945]: W1014 15:17:11.380334 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda154a74e_14ee_43fd_a9f7_e2a8fbffdfe8.slice/crio-14994abe2490ef9a5faaf6d94e55d98a5c8ab89f3539b3f42d137bcd007b837b WatchSource:0}: Error finding container 14994abe2490ef9a5faaf6d94e55d98a5c8ab89f3539b3f42d137bcd007b837b: Status 404 returned error can't find the container with id 14994abe2490ef9a5faaf6d94e55d98a5c8ab89f3539b3f42d137bcd007b837b Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.381443 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.383416 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5lphm"] Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.383810 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.429824 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-secret-volume\") pod \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\" (UID: \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\") " Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.429911 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm7q8\" (UniqueName: \"kubernetes.io/projected/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-kube-api-access-wm7q8\") pod \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\" (UID: \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\") " Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.430004 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-config-volume\") pod \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\" (UID: \"3ec13203-2a26-4de1-b02f-30c58d8ab5a1\") " Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.430197 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73627872-659a-47d9-8aa5-c55682facdc7-utilities\") pod \"redhat-operators-5lphm\" (UID: \"73627872-659a-47d9-8aa5-c55682facdc7\") " pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.430254 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd5kh\" (UniqueName: \"kubernetes.io/projected/73627872-659a-47d9-8aa5-c55682facdc7-kube-api-access-fd5kh\") pod \"redhat-operators-5lphm\" (UID: \"73627872-659a-47d9-8aa5-c55682facdc7\") " pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.430307 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73627872-659a-47d9-8aa5-c55682facdc7-catalog-content\") pod \"redhat-operators-5lphm\" (UID: \"73627872-659a-47d9-8aa5-c55682facdc7\") " pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.431795 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-config-volume" (OuterVolumeSpecName: "config-volume") pod "3ec13203-2a26-4de1-b02f-30c58d8ab5a1" (UID: "3ec13203-2a26-4de1-b02f-30c58d8ab5a1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.435818 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-kube-api-access-wm7q8" (OuterVolumeSpecName: "kube-api-access-wm7q8") pod "3ec13203-2a26-4de1-b02f-30c58d8ab5a1" (UID: "3ec13203-2a26-4de1-b02f-30c58d8ab5a1"). InnerVolumeSpecName "kube-api-access-wm7q8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.437030 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3ec13203-2a26-4de1-b02f-30c58d8ab5a1" (UID: "3ec13203-2a26-4de1-b02f-30c58d8ab5a1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.531556 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73627872-659a-47d9-8aa5-c55682facdc7-catalog-content\") pod \"redhat-operators-5lphm\" (UID: \"73627872-659a-47d9-8aa5-c55682facdc7\") " pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.531650 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73627872-659a-47d9-8aa5-c55682facdc7-utilities\") pod \"redhat-operators-5lphm\" (UID: \"73627872-659a-47d9-8aa5-c55682facdc7\") " pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.531697 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd5kh\" (UniqueName: \"kubernetes.io/projected/73627872-659a-47d9-8aa5-c55682facdc7-kube-api-access-fd5kh\") pod \"redhat-operators-5lphm\" (UID: \"73627872-659a-47d9-8aa5-c55682facdc7\") " pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.531762 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.531778 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.531789 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm7q8\" (UniqueName: \"kubernetes.io/projected/3ec13203-2a26-4de1-b02f-30c58d8ab5a1-kube-api-access-wm7q8\") on node \"crc\" DevicePath \"\"" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.533131 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73627872-659a-47d9-8aa5-c55682facdc7-catalog-content\") pod \"redhat-operators-5lphm\" (UID: \"73627872-659a-47d9-8aa5-c55682facdc7\") " pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.533355 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73627872-659a-47d9-8aa5-c55682facdc7-utilities\") pod \"redhat-operators-5lphm\" (UID: \"73627872-659a-47d9-8aa5-c55682facdc7\") " pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.549152 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd5kh\" (UniqueName: \"kubernetes.io/projected/73627872-659a-47d9-8aa5-c55682facdc7-kube-api-access-fd5kh\") pod \"redhat-operators-5lphm\" (UID: \"73627872-659a-47d9-8aa5-c55682facdc7\") " pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.693502 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.745407 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hrlc6"] Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.757326 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hrlc6"] Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.757417 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.834560 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-utilities\") pod \"redhat-operators-hrlc6\" (UID: \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\") " pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.834596 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-catalog-content\") pod \"redhat-operators-hrlc6\" (UID: \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\") " pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.834633 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kt6g\" (UniqueName: \"kubernetes.io/projected/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-kube-api-access-5kt6g\") pod \"redhat-operators-hrlc6\" (UID: \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\") " pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:17:11 crc kubenswrapper[4945]: E1014 15:17:11.910672 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-pod625cab2a_e7e1_44f7_b06a_61651f5c7d88.slice/crio-conmon-354d3859f29aa75d70b9cf2711b562393f9864d5447d5ea059275964a816e541.scope\": RecentStats: unable to find data in memory cache]" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.936352 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kt6g\" (UniqueName: \"kubernetes.io/projected/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-kube-api-access-5kt6g\") pod \"redhat-operators-hrlc6\" (UID: \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\") " pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.936455 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-utilities\") pod \"redhat-operators-hrlc6\" (UID: \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\") " pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.936476 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-catalog-content\") pod \"redhat-operators-hrlc6\" (UID: \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\") " pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.937218 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-catalog-content\") pod \"redhat-operators-hrlc6\" (UID: \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\") " pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.937831 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-utilities\") pod \"redhat-operators-hrlc6\" (UID: \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\") " pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.959389 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kt6g\" (UniqueName: \"kubernetes.io/projected/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-kube-api-access-5kt6g\") pod \"redhat-operators-hrlc6\" (UID: \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\") " pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:17:11 crc kubenswrapper[4945]: I1014 15:17:11.990738 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5lphm"] Oct 14 15:17:12 crc kubenswrapper[4945]: W1014 15:17:12.018806 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73627872_659a_47d9_8aa5_c55682facdc7.slice/crio-fec1210ef8844855ea194102d51806a3dfb866539957ba3ef118f41202efb2ad WatchSource:0}: Error finding container fec1210ef8844855ea194102d51806a3dfb866539957ba3ef118f41202efb2ad: Status 404 returned error can't find the container with id fec1210ef8844855ea194102d51806a3dfb866539957ba3ef118f41202efb2ad Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.063559 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:12 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:12 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:12 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.063646 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.115646 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lphm" event={"ID":"73627872-659a-47d9-8aa5-c55682facdc7","Type":"ContainerStarted","Data":"fec1210ef8844855ea194102d51806a3dfb866539957ba3ef118f41202efb2ad"} Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.120790 4945 generic.go:334] "Generic (PLEG): container finished" podID="625cab2a-e7e1-44f7-b06a-61651f5c7d88" containerID="354d3859f29aa75d70b9cf2711b562393f9864d5447d5ea059275964a816e541" exitCode=0 Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.120842 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"625cab2a-e7e1-44f7-b06a-61651f5c7d88","Type":"ContainerDied","Data":"354d3859f29aa75d70b9cf2711b562393f9864d5447d5ea059275964a816e541"} Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.125516 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" event={"ID":"3ec13203-2a26-4de1-b02f-30c58d8ab5a1","Type":"ContainerDied","Data":"87556adef92a50d930116aceed4401b8295109ea34de9993164f10e7e97e090e"} Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.125541 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87556adef92a50d930116aceed4401b8295109ea34de9993164f10e7e97e090e" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.125572 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340915-lj4bw" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.145279 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.172702 4945 generic.go:334] "Generic (PLEG): container finished" podID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" containerID="5d97db5d81371a633e2a299581dadae6123f02edb083acbc9f91e5eb5270b394" exitCode=0 Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.172743 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-prqws" event={"ID":"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8","Type":"ContainerDied","Data":"5d97db5d81371a633e2a299581dadae6123f02edb083acbc9f91e5eb5270b394"} Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.172764 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-prqws" event={"ID":"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8","Type":"ContainerStarted","Data":"14994abe2490ef9a5faaf6d94e55d98a5c8ab89f3539b3f42d137bcd007b837b"} Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.326719 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cv87 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.326775 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cv87" podUID="58804eb4-1016-427d-bcf6-fcfa9e8864df" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.331091 4945 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cv87 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.331161 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6cv87" podUID="58804eb4-1016-427d-bcf6-fcfa9e8864df" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.450396 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.453618 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hrlc6"] Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.570184 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.590065 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mnqtf" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.630846 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.636313 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8f2dm" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.893928 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.893959 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.909053 4945 patch_prober.go:28] interesting pod/console-f9d7485db-ccxsw container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 14 15:17:12 crc kubenswrapper[4945]: I1014 15:17:12.909103 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ccxsw" podUID="06ecdccd-2560-48ea-a7f3-87e3e75488a0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.062191 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.065644 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:13 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:13 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:13 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.065687 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.181964 4945 generic.go:334] "Generic (PLEG): container finished" podID="73627872-659a-47d9-8aa5-c55682facdc7" containerID="e19e9407cbcebb564ec22acbbc1d04496e37b5e45fd9fe7cc940c39381292cd5" exitCode=0 Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.183170 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lphm" event={"ID":"73627872-659a-47d9-8aa5-c55682facdc7","Type":"ContainerDied","Data":"e19e9407cbcebb564ec22acbbc1d04496e37b5e45fd9fe7cc940c39381292cd5"} Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.187269 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrlc6" event={"ID":"40554eb6-5285-4ae4-8e2b-ad6f546c0b53","Type":"ContainerDied","Data":"8d76a62a998efc5c5a21bdc496893e1bed4271e56d9e3d35dbbf6d57714e2d46"} Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.187949 4945 generic.go:334] "Generic (PLEG): container finished" podID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" containerID="8d76a62a998efc5c5a21bdc496893e1bed4271e56d9e3d35dbbf6d57714e2d46" exitCode=0 Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.188122 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrlc6" event={"ID":"40554eb6-5285-4ae4-8e2b-ad6f546c0b53","Type":"ContainerStarted","Data":"3a111f584926473bc6adb41511b70d9cd312e9fcd4cc080442cd16d1a7bfff44"} Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.472160 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.581914 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/625cab2a-e7e1-44f7-b06a-61651f5c7d88-kube-api-access\") pod \"625cab2a-e7e1-44f7-b06a-61651f5c7d88\" (UID: \"625cab2a-e7e1-44f7-b06a-61651f5c7d88\") " Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.582065 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/625cab2a-e7e1-44f7-b06a-61651f5c7d88-kubelet-dir\") pod \"625cab2a-e7e1-44f7-b06a-61651f5c7d88\" (UID: \"625cab2a-e7e1-44f7-b06a-61651f5c7d88\") " Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.582366 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/625cab2a-e7e1-44f7-b06a-61651f5c7d88-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "625cab2a-e7e1-44f7-b06a-61651f5c7d88" (UID: "625cab2a-e7e1-44f7-b06a-61651f5c7d88"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.582530 4945 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/625cab2a-e7e1-44f7-b06a-61651f5c7d88-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.588243 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/625cab2a-e7e1-44f7-b06a-61651f5c7d88-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "625cab2a-e7e1-44f7-b06a-61651f5c7d88" (UID: "625cab2a-e7e1-44f7-b06a-61651f5c7d88"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.683413 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/625cab2a-e7e1-44f7-b06a-61651f5c7d88-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.797077 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-l72k8" Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.988425 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:17:13 crc kubenswrapper[4945]: I1014 15:17:13.993519 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be17bc84-909e-4150-8ee7-9a7d23173ba0-metrics-certs\") pod \"network-metrics-daemon-ftg5h\" (UID: \"be17bc84-909e-4150-8ee7-9a7d23173ba0\") " pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.062122 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:14 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:14 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:14 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.062200 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.133074 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ftg5h" Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.229854 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"625cab2a-e7e1-44f7-b06a-61651f5c7d88","Type":"ContainerDied","Data":"bc5c3d0f43b5ecb81b5dab38a1b01943a7a0a692e0e486d0d3cff7bd1edfc6cb"} Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.229907 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc5c3d0f43b5ecb81b5dab38a1b01943a7a0a692e0e486d0d3cff7bd1edfc6cb" Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.229977 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.489770 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-ftg5h"] Oct 14 15:17:14 crc kubenswrapper[4945]: W1014 15:17:14.524865 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe17bc84_909e_4150_8ee7_9a7d23173ba0.slice/crio-9d0d5b4d473ac3e82ed043733f584f31d897a6285bd1400bddeddaec8f0d6c4c WatchSource:0}: Error finding container 9d0d5b4d473ac3e82ed043733f584f31d897a6285bd1400bddeddaec8f0d6c4c: Status 404 returned error can't find the container with id 9d0d5b4d473ac3e82ed043733f584f31d897a6285bd1400bddeddaec8f0d6c4c Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.869263 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 14 15:17:14 crc kubenswrapper[4945]: E1014 15:17:14.869473 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="625cab2a-e7e1-44f7-b06a-61651f5c7d88" containerName="pruner" Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.869485 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="625cab2a-e7e1-44f7-b06a-61651f5c7d88" containerName="pruner" Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.869588 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="625cab2a-e7e1-44f7-b06a-61651f5c7d88" containerName="pruner" Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.869951 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.871527 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.872281 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 14 15:17:14 crc kubenswrapper[4945]: I1014 15:17:14.879477 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 14 15:17:15 crc kubenswrapper[4945]: I1014 15:17:15.001235 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc1d96aa-1a35-45ee-99f9-2f28440fa345-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"dc1d96aa-1a35-45ee-99f9-2f28440fa345\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 15:17:15 crc kubenswrapper[4945]: I1014 15:17:15.001282 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc1d96aa-1a35-45ee-99f9-2f28440fa345-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"dc1d96aa-1a35-45ee-99f9-2f28440fa345\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 15:17:15 crc kubenswrapper[4945]: I1014 15:17:15.063171 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:15 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:15 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:15 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:15 crc kubenswrapper[4945]: I1014 15:17:15.063249 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:15 crc kubenswrapper[4945]: I1014 15:17:15.102381 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc1d96aa-1a35-45ee-99f9-2f28440fa345-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"dc1d96aa-1a35-45ee-99f9-2f28440fa345\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 15:17:15 crc kubenswrapper[4945]: I1014 15:17:15.102434 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc1d96aa-1a35-45ee-99f9-2f28440fa345-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"dc1d96aa-1a35-45ee-99f9-2f28440fa345\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 15:17:15 crc kubenswrapper[4945]: I1014 15:17:15.102900 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc1d96aa-1a35-45ee-99f9-2f28440fa345-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"dc1d96aa-1a35-45ee-99f9-2f28440fa345\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 15:17:15 crc kubenswrapper[4945]: I1014 15:17:15.119327 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc1d96aa-1a35-45ee-99f9-2f28440fa345-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"dc1d96aa-1a35-45ee-99f9-2f28440fa345\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 15:17:15 crc kubenswrapper[4945]: I1014 15:17:15.210398 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 15:17:15 crc kubenswrapper[4945]: I1014 15:17:15.247661 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" event={"ID":"be17bc84-909e-4150-8ee7-9a7d23173ba0","Type":"ContainerStarted","Data":"a7fb1b5181f54f3ad6b1a3d2bb3be51694f8b1cb5a0327fe67bec536fc81a15a"} Oct 14 15:17:15 crc kubenswrapper[4945]: I1014 15:17:15.247723 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" event={"ID":"be17bc84-909e-4150-8ee7-9a7d23173ba0","Type":"ContainerStarted","Data":"9d0d5b4d473ac3e82ed043733f584f31d897a6285bd1400bddeddaec8f0d6c4c"} Oct 14 15:17:15 crc kubenswrapper[4945]: I1014 15:17:15.529426 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 14 15:17:16 crc kubenswrapper[4945]: I1014 15:17:16.064221 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:16 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:16 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:16 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:16 crc kubenswrapper[4945]: I1014 15:17:16.064671 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:16 crc kubenswrapper[4945]: I1014 15:17:16.261191 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"dc1d96aa-1a35-45ee-99f9-2f28440fa345","Type":"ContainerStarted","Data":"cbd7d4ac83f8471162eeca94bd2edcdde05bd1ba515bbe287b0d097d5579f7a1"} Oct 14 15:17:16 crc kubenswrapper[4945]: I1014 15:17:16.799075 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:17:16 crc kubenswrapper[4945]: I1014 15:17:16.799382 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:17:17 crc kubenswrapper[4945]: I1014 15:17:17.061904 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:17 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:17 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:17 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:17 crc kubenswrapper[4945]: I1014 15:17:17.061955 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:17 crc kubenswrapper[4945]: I1014 15:17:17.268687 4945 generic.go:334] "Generic (PLEG): container finished" podID="dc1d96aa-1a35-45ee-99f9-2f28440fa345" containerID="0ad2ecbd02e8ec174b511a7df0fc5407a0f30027598725666cbee5e293401ed0" exitCode=0 Oct 14 15:17:17 crc kubenswrapper[4945]: I1014 15:17:17.268752 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"dc1d96aa-1a35-45ee-99f9-2f28440fa345","Type":"ContainerDied","Data":"0ad2ecbd02e8ec174b511a7df0fc5407a0f30027598725666cbee5e293401ed0"} Oct 14 15:17:17 crc kubenswrapper[4945]: I1014 15:17:17.277176 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ftg5h" event={"ID":"be17bc84-909e-4150-8ee7-9a7d23173ba0","Type":"ContainerStarted","Data":"2ee8387dd5e3dc37452e5824de7f1f9c4f3b970f7857af8042fb105cd7de0de3"} Oct 14 15:17:17 crc kubenswrapper[4945]: I1014 15:17:17.293225 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-ftg5h" podStartSLOduration=146.293210051 podStartE2EDuration="2m26.293210051s" podCreationTimestamp="2025-10-14 15:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:17:17.291973756 +0000 UTC m=+167.276022114" watchObservedRunningTime="2025-10-14 15:17:17.293210051 +0000 UTC m=+167.277258419" Oct 14 15:17:18 crc kubenswrapper[4945]: I1014 15:17:18.062868 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:18 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:18 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:18 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:18 crc kubenswrapper[4945]: I1014 15:17:18.062936 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:19 crc kubenswrapper[4945]: I1014 15:17:19.061899 4945 patch_prober.go:28] interesting pod/router-default-5444994796-6j5pp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 15:17:19 crc kubenswrapper[4945]: [-]has-synced failed: reason withheld Oct 14 15:17:19 crc kubenswrapper[4945]: [+]process-running ok Oct 14 15:17:19 crc kubenswrapper[4945]: healthz check failed Oct 14 15:17:19 crc kubenswrapper[4945]: I1014 15:17:19.062234 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6j5pp" podUID="d161354c-4c64-4c8f-a368-3462e7d3c5ac" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 15:17:20 crc kubenswrapper[4945]: I1014 15:17:20.063053 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:20 crc kubenswrapper[4945]: I1014 15:17:20.064964 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-6j5pp" Oct 14 15:17:22 crc kubenswrapper[4945]: I1014 15:17:22.334702 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-6cv87" Oct 14 15:17:22 crc kubenswrapper[4945]: I1014 15:17:22.904741 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:22 crc kubenswrapper[4945]: I1014 15:17:22.914289 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:17:24 crc kubenswrapper[4945]: I1014 15:17:24.568197 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 15:17:24 crc kubenswrapper[4945]: I1014 15:17:24.661022 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc1d96aa-1a35-45ee-99f9-2f28440fa345-kube-api-access\") pod \"dc1d96aa-1a35-45ee-99f9-2f28440fa345\" (UID: \"dc1d96aa-1a35-45ee-99f9-2f28440fa345\") " Oct 14 15:17:24 crc kubenswrapper[4945]: I1014 15:17:24.661238 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc1d96aa-1a35-45ee-99f9-2f28440fa345-kubelet-dir\") pod \"dc1d96aa-1a35-45ee-99f9-2f28440fa345\" (UID: \"dc1d96aa-1a35-45ee-99f9-2f28440fa345\") " Oct 14 15:17:24 crc kubenswrapper[4945]: I1014 15:17:24.661344 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc1d96aa-1a35-45ee-99f9-2f28440fa345-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "dc1d96aa-1a35-45ee-99f9-2f28440fa345" (UID: "dc1d96aa-1a35-45ee-99f9-2f28440fa345"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:17:24 crc kubenswrapper[4945]: I1014 15:17:24.661632 4945 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc1d96aa-1a35-45ee-99f9-2f28440fa345-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 14 15:17:24 crc kubenswrapper[4945]: I1014 15:17:24.666457 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc1d96aa-1a35-45ee-99f9-2f28440fa345-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "dc1d96aa-1a35-45ee-99f9-2f28440fa345" (UID: "dc1d96aa-1a35-45ee-99f9-2f28440fa345"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:17:24 crc kubenswrapper[4945]: I1014 15:17:24.763222 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc1d96aa-1a35-45ee-99f9-2f28440fa345-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 15:17:25 crc kubenswrapper[4945]: I1014 15:17:25.333728 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"dc1d96aa-1a35-45ee-99f9-2f28440fa345","Type":"ContainerDied","Data":"cbd7d4ac83f8471162eeca94bd2edcdde05bd1ba515bbe287b0d097d5579f7a1"} Oct 14 15:17:25 crc kubenswrapper[4945]: I1014 15:17:25.333782 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbd7d4ac83f8471162eeca94bd2edcdde05bd1ba515bbe287b0d097d5579f7a1" Oct 14 15:17:25 crc kubenswrapper[4945]: I1014 15:17:25.333845 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 15:17:29 crc kubenswrapper[4945]: I1014 15:17:29.259925 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:17:39 crc kubenswrapper[4945]: I1014 15:17:39.359738 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 15:17:40 crc kubenswrapper[4945]: E1014 15:17:40.080485 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 14 15:17:40 crc kubenswrapper[4945]: E1014 15:17:40.080722 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-klwwd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tfj4f_openshift-marketplace(ec783233-16ca-4e17-8923-71455b2056d7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 15:17:40 crc kubenswrapper[4945]: E1014 15:17:40.082005 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-tfj4f" podUID="ec783233-16ca-4e17-8923-71455b2056d7" Oct 14 15:17:41 crc kubenswrapper[4945]: E1014 15:17:41.374679 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tfj4f" podUID="ec783233-16ca-4e17-8923-71455b2056d7" Oct 14 15:17:42 crc kubenswrapper[4945]: I1014 15:17:42.758716 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8rkjt" Oct 14 15:17:44 crc kubenswrapper[4945]: E1014 15:17:44.611584 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 14 15:17:44 crc kubenswrapper[4945]: E1014 15:17:44.612044 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lkwqk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vs4wr_openshift-marketplace(5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 15:17:44 crc kubenswrapper[4945]: E1014 15:17:44.613320 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vs4wr" podUID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" Oct 14 15:17:46 crc kubenswrapper[4945]: E1014 15:17:46.644616 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vs4wr" podUID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" Oct 14 15:17:46 crc kubenswrapper[4945]: I1014 15:17:46.796339 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:17:46 crc kubenswrapper[4945]: I1014 15:17:46.796417 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:17:51 crc kubenswrapper[4945]: E1014 15:17:51.738536 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 14 15:17:51 crc kubenswrapper[4945]: E1014 15:17:51.740174 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5kt6g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-hrlc6_openshift-marketplace(40554eb6-5285-4ae4-8e2b-ad6f546c0b53): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 15:17:51 crc kubenswrapper[4945]: E1014 15:17:51.741698 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-hrlc6" podUID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" Oct 14 15:17:52 crc kubenswrapper[4945]: E1014 15:17:52.486399 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-hrlc6" podUID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" Oct 14 15:17:55 crc kubenswrapper[4945]: E1014 15:17:55.643729 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 14 15:17:55 crc kubenswrapper[4945]: E1014 15:17:55.644328 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4mm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-b27lt_openshift-marketplace(27eaa4e2-7ffd-41b3-9d3d-418b43669de8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 15:17:55 crc kubenswrapper[4945]: E1014 15:17:55.645695 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-b27lt" podUID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" Oct 14 15:17:55 crc kubenswrapper[4945]: E1014 15:17:55.679391 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 14 15:17:55 crc kubenswrapper[4945]: E1014 15:17:55.679686 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j9gzn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-prqws_openshift-marketplace(a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 15:17:55 crc kubenswrapper[4945]: E1014 15:17:55.680959 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-prqws" podUID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" Oct 14 15:17:55 crc kubenswrapper[4945]: E1014 15:17:55.683263 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 14 15:17:55 crc kubenswrapper[4945]: E1014 15:17:55.683504 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jm4j2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-wvfqb_openshift-marketplace(2424e499-800f-481e-95e2-84257025d221): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 15:17:55 crc kubenswrapper[4945]: E1014 15:17:55.686175 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-wvfqb" podUID="2424e499-800f-481e-95e2-84257025d221" Oct 14 15:17:55 crc kubenswrapper[4945]: E1014 15:17:55.690146 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 14 15:17:55 crc kubenswrapper[4945]: E1014 15:17:55.690395 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tmnsl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-8z5p9_openshift-marketplace(ba2d8579-0ecb-46a2-950b-c48d660ed7ab): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 15:17:55 crc kubenswrapper[4945]: E1014 15:17:55.691699 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-8z5p9" podUID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" Oct 14 15:17:56 crc kubenswrapper[4945]: I1014 15:17:56.507421 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lphm" event={"ID":"73627872-659a-47d9-8aa5-c55682facdc7","Type":"ContainerStarted","Data":"38084c83fb9d2954669b62c23f53d4ac610dcae410397f17a35896126405f0b9"} Oct 14 15:17:56 crc kubenswrapper[4945]: E1014 15:17:56.510077 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-8z5p9" podUID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" Oct 14 15:17:56 crc kubenswrapper[4945]: E1014 15:17:56.510312 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-b27lt" podUID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" Oct 14 15:17:56 crc kubenswrapper[4945]: E1014 15:17:56.510077 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-wvfqb" podUID="2424e499-800f-481e-95e2-84257025d221" Oct 14 15:17:56 crc kubenswrapper[4945]: E1014 15:17:56.510802 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-prqws" podUID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" Oct 14 15:17:57 crc kubenswrapper[4945]: I1014 15:17:57.514581 4945 generic.go:334] "Generic (PLEG): container finished" podID="73627872-659a-47d9-8aa5-c55682facdc7" containerID="38084c83fb9d2954669b62c23f53d4ac610dcae410397f17a35896126405f0b9" exitCode=0 Oct 14 15:17:57 crc kubenswrapper[4945]: I1014 15:17:57.515242 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lphm" event={"ID":"73627872-659a-47d9-8aa5-c55682facdc7","Type":"ContainerDied","Data":"38084c83fb9d2954669b62c23f53d4ac610dcae410397f17a35896126405f0b9"} Oct 14 15:17:58 crc kubenswrapper[4945]: I1014 15:17:58.523855 4945 generic.go:334] "Generic (PLEG): container finished" podID="ec783233-16ca-4e17-8923-71455b2056d7" containerID="7bf4a9f23db2d0f71bc5cd308b83d01cb61b4d2a57986c153bc5f6e1b67bc851" exitCode=0 Oct 14 15:17:58 crc kubenswrapper[4945]: I1014 15:17:58.523932 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfj4f" event={"ID":"ec783233-16ca-4e17-8923-71455b2056d7","Type":"ContainerDied","Data":"7bf4a9f23db2d0f71bc5cd308b83d01cb61b4d2a57986c153bc5f6e1b67bc851"} Oct 14 15:17:59 crc kubenswrapper[4945]: I1014 15:17:59.538303 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lphm" event={"ID":"73627872-659a-47d9-8aa5-c55682facdc7","Type":"ContainerStarted","Data":"1411170f7c2bd1ae8ec61f07c42baa61524c6affad3a7ec464f6240554d8edc7"} Oct 14 15:17:59 crc kubenswrapper[4945]: I1014 15:17:59.566863 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5lphm" podStartSLOduration=3.265634738 podStartE2EDuration="48.566840269s" podCreationTimestamp="2025-10-14 15:17:11 +0000 UTC" firstStartedPulling="2025-10-14 15:17:13.185146463 +0000 UTC m=+163.169194832" lastFinishedPulling="2025-10-14 15:17:58.486352005 +0000 UTC m=+208.470400363" observedRunningTime="2025-10-14 15:17:59.565033342 +0000 UTC m=+209.549081760" watchObservedRunningTime="2025-10-14 15:17:59.566840269 +0000 UTC m=+209.550888647" Oct 14 15:18:01 crc kubenswrapper[4945]: I1014 15:18:01.694099 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:18:01 crc kubenswrapper[4945]: I1014 15:18:01.694634 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:18:02 crc kubenswrapper[4945]: I1014 15:18:02.557095 4945 generic.go:334] "Generic (PLEG): container finished" podID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" containerID="f4e648faedb9ce11a928967ed75d54653e1567161c49af283f14f824ecc991d4" exitCode=0 Oct 14 15:18:02 crc kubenswrapper[4945]: I1014 15:18:02.557221 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vs4wr" event={"ID":"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039","Type":"ContainerDied","Data":"f4e648faedb9ce11a928967ed75d54653e1567161c49af283f14f824ecc991d4"} Oct 14 15:18:02 crc kubenswrapper[4945]: I1014 15:18:02.560952 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfj4f" event={"ID":"ec783233-16ca-4e17-8923-71455b2056d7","Type":"ContainerStarted","Data":"2c274dd86cc88755f803c9de36dbd4388371808e03e060a0d26f6af565114ccf"} Oct 14 15:18:02 crc kubenswrapper[4945]: I1014 15:18:02.974510 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5lphm" podUID="73627872-659a-47d9-8aa5-c55682facdc7" containerName="registry-server" probeResult="failure" output=< Oct 14 15:18:02 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 14 15:18:02 crc kubenswrapper[4945]: > Oct 14 15:18:05 crc kubenswrapper[4945]: I1014 15:18:05.785279 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tfj4f" podStartSLOduration=6.51918083 podStartE2EDuration="57.785256082s" podCreationTimestamp="2025-10-14 15:17:08 +0000 UTC" firstStartedPulling="2025-10-14 15:17:10.07795282 +0000 UTC m=+160.062001188" lastFinishedPulling="2025-10-14 15:18:01.344028062 +0000 UTC m=+211.328076440" observedRunningTime="2025-10-14 15:18:02.605463213 +0000 UTC m=+212.589511581" watchObservedRunningTime="2025-10-14 15:18:05.785256082 +0000 UTC m=+215.769304450" Oct 14 15:18:06 crc kubenswrapper[4945]: I1014 15:18:06.587393 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vs4wr" event={"ID":"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039","Type":"ContainerStarted","Data":"97e4d44348b3e6dea11ce84bee2e6afadb6b6534f3e24235f30d7c34fc04f864"} Oct 14 15:18:06 crc kubenswrapper[4945]: I1014 15:18:06.617687 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vs4wr" podStartSLOduration=2.817685754 podStartE2EDuration="58.617660261s" podCreationTimestamp="2025-10-14 15:17:08 +0000 UTC" firstStartedPulling="2025-10-14 15:17:10.088622701 +0000 UTC m=+160.072671069" lastFinishedPulling="2025-10-14 15:18:05.888597158 +0000 UTC m=+215.872645576" observedRunningTime="2025-10-14 15:18:06.61479588 +0000 UTC m=+216.598844258" watchObservedRunningTime="2025-10-14 15:18:06.617660261 +0000 UTC m=+216.601708659" Oct 14 15:18:08 crc kubenswrapper[4945]: I1014 15:18:08.485933 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:18:08 crc kubenswrapper[4945]: I1014 15:18:08.486314 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:18:08 crc kubenswrapper[4945]: I1014 15:18:08.600325 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrlc6" event={"ID":"40554eb6-5285-4ae4-8e2b-ad6f546c0b53","Type":"ContainerStarted","Data":"c4d0ef4527004805103029e330db7e7010e0782dd9ee2c747cd75fea11102723"} Oct 14 15:18:08 crc kubenswrapper[4945]: I1014 15:18:08.739134 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:18:08 crc kubenswrapper[4945]: I1014 15:18:08.739473 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:18:08 crc kubenswrapper[4945]: I1014 15:18:08.799300 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:18:09 crc kubenswrapper[4945]: I1014 15:18:09.541708 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-vs4wr" podUID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" containerName="registry-server" probeResult="failure" output=< Oct 14 15:18:09 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 14 15:18:09 crc kubenswrapper[4945]: > Oct 14 15:18:09 crc kubenswrapper[4945]: I1014 15:18:09.610995 4945 generic.go:334] "Generic (PLEG): container finished" podID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" containerID="c4d0ef4527004805103029e330db7e7010e0782dd9ee2c747cd75fea11102723" exitCode=0 Oct 14 15:18:09 crc kubenswrapper[4945]: I1014 15:18:09.611037 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrlc6" event={"ID":"40554eb6-5285-4ae4-8e2b-ad6f546c0b53","Type":"ContainerDied","Data":"c4d0ef4527004805103029e330db7e7010e0782dd9ee2c747cd75fea11102723"} Oct 14 15:18:09 crc kubenswrapper[4945]: I1014 15:18:09.657628 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:18:11 crc kubenswrapper[4945]: I1014 15:18:11.751160 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:18:11 crc kubenswrapper[4945]: I1014 15:18:11.806034 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:18:16 crc kubenswrapper[4945]: I1014 15:18:16.796216 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:18:16 crc kubenswrapper[4945]: I1014 15:18:16.796505 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:18:16 crc kubenswrapper[4945]: I1014 15:18:16.796551 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:18:16 crc kubenswrapper[4945]: I1014 15:18:16.797091 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd"} pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 15:18:16 crc kubenswrapper[4945]: I1014 15:18:16.797190 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" containerID="cri-o://e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd" gracePeriod=600 Oct 14 15:18:17 crc kubenswrapper[4945]: I1014 15:18:17.661560 4945 generic.go:334] "Generic (PLEG): container finished" podID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerID="e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd" exitCode=0 Oct 14 15:18:17 crc kubenswrapper[4945]: I1014 15:18:17.661685 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerDied","Data":"e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd"} Oct 14 15:18:18 crc kubenswrapper[4945]: I1014 15:18:18.562641 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:18:18 crc kubenswrapper[4945]: I1014 15:18:18.640366 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:18:21 crc kubenswrapper[4945]: I1014 15:18:21.689435 4945 generic.go:334] "Generic (PLEG): container finished" podID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" containerID="df23c4e4922478bda19b51688e38fd7314803f26cae413bfbbd6ff0a33ad765f" exitCode=0 Oct 14 15:18:21 crc kubenswrapper[4945]: I1014 15:18:21.689502 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z5p9" event={"ID":"ba2d8579-0ecb-46a2-950b-c48d660ed7ab","Type":"ContainerDied","Data":"df23c4e4922478bda19b51688e38fd7314803f26cae413bfbbd6ff0a33ad765f"} Oct 14 15:18:21 crc kubenswrapper[4945]: I1014 15:18:21.694901 4945 generic.go:334] "Generic (PLEG): container finished" podID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" containerID="c73db62ba2cebd6ed88186390619656b58ebbffbc1154ab7a657d3b591f51a23" exitCode=0 Oct 14 15:18:21 crc kubenswrapper[4945]: I1014 15:18:21.694945 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b27lt" event={"ID":"27eaa4e2-7ffd-41b3-9d3d-418b43669de8","Type":"ContainerDied","Data":"c73db62ba2cebd6ed88186390619656b58ebbffbc1154ab7a657d3b591f51a23"} Oct 14 15:18:21 crc kubenswrapper[4945]: I1014 15:18:21.701679 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerStarted","Data":"c8c34e87ab2d9cc333cc18624592bb9519696247cdd2c0458ada848829da5a55"} Oct 14 15:18:21 crc kubenswrapper[4945]: I1014 15:18:21.705353 4945 generic.go:334] "Generic (PLEG): container finished" podID="2424e499-800f-481e-95e2-84257025d221" containerID="f4d5b80642a6cfbc6c43ebeb0909605ed632b15c25c51fea5b78ed506ab9cb71" exitCode=0 Oct 14 15:18:21 crc kubenswrapper[4945]: I1014 15:18:21.705438 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvfqb" event={"ID":"2424e499-800f-481e-95e2-84257025d221","Type":"ContainerDied","Data":"f4d5b80642a6cfbc6c43ebeb0909605ed632b15c25c51fea5b78ed506ab9cb71"} Oct 14 15:18:21 crc kubenswrapper[4945]: I1014 15:18:21.711310 4945 generic.go:334] "Generic (PLEG): container finished" podID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" containerID="98229a8761089893f1ece194ecb43cd2702c5209c221caadac1c55d1356c8704" exitCode=0 Oct 14 15:18:21 crc kubenswrapper[4945]: I1014 15:18:21.711347 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-prqws" event={"ID":"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8","Type":"ContainerDied","Data":"98229a8761089893f1ece194ecb43cd2702c5209c221caadac1c55d1356c8704"} Oct 14 15:18:21 crc kubenswrapper[4945]: I1014 15:18:21.717979 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrlc6" event={"ID":"40554eb6-5285-4ae4-8e2b-ad6f546c0b53","Type":"ContainerStarted","Data":"bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b"} Oct 14 15:18:21 crc kubenswrapper[4945]: I1014 15:18:21.817845 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hrlc6" podStartSLOduration=4.131666774 podStartE2EDuration="1m10.817825701s" podCreationTimestamp="2025-10-14 15:17:11 +0000 UTC" firstStartedPulling="2025-10-14 15:17:13.190326409 +0000 UTC m=+163.174374777" lastFinishedPulling="2025-10-14 15:18:19.876485336 +0000 UTC m=+229.860533704" observedRunningTime="2025-10-14 15:18:21.815378063 +0000 UTC m=+231.799426471" watchObservedRunningTime="2025-10-14 15:18:21.817825701 +0000 UTC m=+231.801874059" Oct 14 15:18:22 crc kubenswrapper[4945]: I1014 15:18:22.146356 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:18:22 crc kubenswrapper[4945]: I1014 15:18:22.146677 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:18:22 crc kubenswrapper[4945]: I1014 15:18:22.724703 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvfqb" event={"ID":"2424e499-800f-481e-95e2-84257025d221","Type":"ContainerStarted","Data":"e4f1e4f5e2dfa4ab90e73fd43ae6ddc896f631a210dc262517aa84beef44203a"} Oct 14 15:18:22 crc kubenswrapper[4945]: I1014 15:18:22.727746 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b27lt" event={"ID":"27eaa4e2-7ffd-41b3-9d3d-418b43669de8","Type":"ContainerStarted","Data":"2278bb953c1e948d862bb06a80c4c88b3b14f37b9785912dfea06a88d9eb31d9"} Oct 14 15:18:22 crc kubenswrapper[4945]: I1014 15:18:22.775773 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wvfqb" podStartSLOduration=2.388675245 podStartE2EDuration="1m14.77575297s" podCreationTimestamp="2025-10-14 15:17:08 +0000 UTC" firstStartedPulling="2025-10-14 15:17:10.085260996 +0000 UTC m=+160.069309364" lastFinishedPulling="2025-10-14 15:18:22.472338721 +0000 UTC m=+232.456387089" observedRunningTime="2025-10-14 15:18:22.757860163 +0000 UTC m=+232.741908551" watchObservedRunningTime="2025-10-14 15:18:22.77575297 +0000 UTC m=+232.759801338" Oct 14 15:18:22 crc kubenswrapper[4945]: I1014 15:18:22.776264 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b27lt" podStartSLOduration=1.540270621 podStartE2EDuration="1m12.776260506s" podCreationTimestamp="2025-10-14 15:17:10 +0000 UTC" firstStartedPulling="2025-10-14 15:17:11.096218713 +0000 UTC m=+161.080267081" lastFinishedPulling="2025-10-14 15:18:22.332208598 +0000 UTC m=+232.316256966" observedRunningTime="2025-10-14 15:18:22.775243864 +0000 UTC m=+232.759292262" watchObservedRunningTime="2025-10-14 15:18:22.776260506 +0000 UTC m=+232.760308874" Oct 14 15:18:23 crc kubenswrapper[4945]: I1014 15:18:23.197429 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hrlc6" podUID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" containerName="registry-server" probeResult="failure" output=< Oct 14 15:18:23 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 14 15:18:23 crc kubenswrapper[4945]: > Oct 14 15:18:23 crc kubenswrapper[4945]: I1014 15:18:23.735071 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z5p9" event={"ID":"ba2d8579-0ecb-46a2-950b-c48d660ed7ab","Type":"ContainerStarted","Data":"050f6964e5cd767cf79a2ecd054c442d972a75a8c22c0e41a23da6992832e53e"} Oct 14 15:18:23 crc kubenswrapper[4945]: I1014 15:18:23.739236 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-prqws" event={"ID":"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8","Type":"ContainerStarted","Data":"e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3"} Oct 14 15:18:23 crc kubenswrapper[4945]: I1014 15:18:23.754836 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8z5p9" podStartSLOduration=3.1838810459999998 podStartE2EDuration="1m15.75482213s" podCreationTimestamp="2025-10-14 15:17:08 +0000 UTC" firstStartedPulling="2025-10-14 15:17:10.076727246 +0000 UTC m=+160.060775614" lastFinishedPulling="2025-10-14 15:18:22.64766831 +0000 UTC m=+232.631716698" observedRunningTime="2025-10-14 15:18:23.753187798 +0000 UTC m=+233.737236166" watchObservedRunningTime="2025-10-14 15:18:23.75482213 +0000 UTC m=+233.738870498" Oct 14 15:18:23 crc kubenswrapper[4945]: I1014 15:18:23.777733 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-prqws" podStartSLOduration=3.131310806 podStartE2EDuration="1m13.777718616s" podCreationTimestamp="2025-10-14 15:17:10 +0000 UTC" firstStartedPulling="2025-10-14 15:17:12.212227338 +0000 UTC m=+162.196275706" lastFinishedPulling="2025-10-14 15:18:22.858635128 +0000 UTC m=+232.842683516" observedRunningTime="2025-10-14 15:18:23.774048729 +0000 UTC m=+233.758097097" watchObservedRunningTime="2025-10-14 15:18:23.777718616 +0000 UTC m=+233.761766984" Oct 14 15:18:28 crc kubenswrapper[4945]: I1014 15:18:28.866815 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:18:28 crc kubenswrapper[4945]: I1014 15:18:28.867307 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:18:28 crc kubenswrapper[4945]: I1014 15:18:28.924637 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:18:29 crc kubenswrapper[4945]: I1014 15:18:29.107280 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:18:29 crc kubenswrapper[4945]: I1014 15:18:29.107333 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:18:29 crc kubenswrapper[4945]: I1014 15:18:29.162637 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:18:29 crc kubenswrapper[4945]: I1014 15:18:29.832333 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:18:29 crc kubenswrapper[4945]: I1014 15:18:29.835162 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:18:30 crc kubenswrapper[4945]: I1014 15:18:30.667287 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:18:30 crc kubenswrapper[4945]: I1014 15:18:30.667569 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:18:30 crc kubenswrapper[4945]: I1014 15:18:30.702102 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:18:30 crc kubenswrapper[4945]: I1014 15:18:30.825385 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:18:31 crc kubenswrapper[4945]: I1014 15:18:31.062493 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:18:31 crc kubenswrapper[4945]: I1014 15:18:31.062845 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:18:31 crc kubenswrapper[4945]: I1014 15:18:31.100386 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:18:31 crc kubenswrapper[4945]: I1014 15:18:31.844772 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:18:32 crc kubenswrapper[4945]: I1014 15:18:32.003223 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wvfqb"] Oct 14 15:18:32 crc kubenswrapper[4945]: I1014 15:18:32.003484 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wvfqb" podUID="2424e499-800f-481e-95e2-84257025d221" containerName="registry-server" containerID="cri-o://e4f1e4f5e2dfa4ab90e73fd43ae6ddc896f631a210dc262517aa84beef44203a" gracePeriod=2 Oct 14 15:18:32 crc kubenswrapper[4945]: I1014 15:18:32.202058 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:18:32 crc kubenswrapper[4945]: I1014 15:18:32.239285 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:18:33 crc kubenswrapper[4945]: I1014 15:18:33.801825 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hrlc6"] Oct 14 15:18:33 crc kubenswrapper[4945]: I1014 15:18:33.805526 4945 generic.go:334] "Generic (PLEG): container finished" podID="2424e499-800f-481e-95e2-84257025d221" containerID="e4f1e4f5e2dfa4ab90e73fd43ae6ddc896f631a210dc262517aa84beef44203a" exitCode=0 Oct 14 15:18:33 crc kubenswrapper[4945]: I1014 15:18:33.805959 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvfqb" event={"ID":"2424e499-800f-481e-95e2-84257025d221","Type":"ContainerDied","Data":"e4f1e4f5e2dfa4ab90e73fd43ae6ddc896f631a210dc262517aa84beef44203a"} Oct 14 15:18:33 crc kubenswrapper[4945]: I1014 15:18:33.806204 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hrlc6" podUID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" containerName="registry-server" containerID="cri-o://bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b" gracePeriod=2 Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.003655 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.170561 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.184632 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2424e499-800f-481e-95e2-84257025d221-catalog-content\") pod \"2424e499-800f-481e-95e2-84257025d221\" (UID: \"2424e499-800f-481e-95e2-84257025d221\") " Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.184809 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm4j2\" (UniqueName: \"kubernetes.io/projected/2424e499-800f-481e-95e2-84257025d221-kube-api-access-jm4j2\") pod \"2424e499-800f-481e-95e2-84257025d221\" (UID: \"2424e499-800f-481e-95e2-84257025d221\") " Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.184850 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2424e499-800f-481e-95e2-84257025d221-utilities\") pod \"2424e499-800f-481e-95e2-84257025d221\" (UID: \"2424e499-800f-481e-95e2-84257025d221\") " Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.185535 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2424e499-800f-481e-95e2-84257025d221-utilities" (OuterVolumeSpecName: "utilities") pod "2424e499-800f-481e-95e2-84257025d221" (UID: "2424e499-800f-481e-95e2-84257025d221"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.192074 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2424e499-800f-481e-95e2-84257025d221-kube-api-access-jm4j2" (OuterVolumeSpecName: "kube-api-access-jm4j2") pod "2424e499-800f-481e-95e2-84257025d221" (UID: "2424e499-800f-481e-95e2-84257025d221"). InnerVolumeSpecName "kube-api-access-jm4j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.286162 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-utilities\") pod \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\" (UID: \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\") " Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.286266 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-catalog-content\") pod \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\" (UID: \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\") " Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.286402 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kt6g\" (UniqueName: \"kubernetes.io/projected/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-kube-api-access-5kt6g\") pod \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\" (UID: \"40554eb6-5285-4ae4-8e2b-ad6f546c0b53\") " Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.286763 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2424e499-800f-481e-95e2-84257025d221-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.286793 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm4j2\" (UniqueName: \"kubernetes.io/projected/2424e499-800f-481e-95e2-84257025d221-kube-api-access-jm4j2\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.286864 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-utilities" (OuterVolumeSpecName: "utilities") pod "40554eb6-5285-4ae4-8e2b-ad6f546c0b53" (UID: "40554eb6-5285-4ae4-8e2b-ad6f546c0b53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.289789 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-kube-api-access-5kt6g" (OuterVolumeSpecName: "kube-api-access-5kt6g") pod "40554eb6-5285-4ae4-8e2b-ad6f546c0b53" (UID: "40554eb6-5285-4ae4-8e2b-ad6f546c0b53"). InnerVolumeSpecName "kube-api-access-5kt6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.387488 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.387518 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kt6g\" (UniqueName: \"kubernetes.io/projected/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-kube-api-access-5kt6g\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.401732 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8z5p9"] Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.401960 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8z5p9" podUID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" containerName="registry-server" containerID="cri-o://050f6964e5cd767cf79a2ecd054c442d972a75a8c22c0e41a23da6992832e53e" gracePeriod=2 Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.415528 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40554eb6-5285-4ae4-8e2b-ad6f546c0b53" (UID: "40554eb6-5285-4ae4-8e2b-ad6f546c0b53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.489294 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40554eb6-5285-4ae4-8e2b-ad6f546c0b53-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.811987 4945 generic.go:334] "Generic (PLEG): container finished" podID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" containerID="bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b" exitCode=0 Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.812088 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrlc6" event={"ID":"40554eb6-5285-4ae4-8e2b-ad6f546c0b53","Type":"ContainerDied","Data":"bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b"} Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.812094 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrlc6" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.812120 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrlc6" event={"ID":"40554eb6-5285-4ae4-8e2b-ad6f546c0b53","Type":"ContainerDied","Data":"3a111f584926473bc6adb41511b70d9cd312e9fcd4cc080442cd16d1a7bfff44"} Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.812145 4945 scope.go:117] "RemoveContainer" containerID="bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.819300 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2424e499-800f-481e-95e2-84257025d221-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2424e499-800f-481e-95e2-84257025d221" (UID: "2424e499-800f-481e-95e2-84257025d221"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.821118 4945 generic.go:334] "Generic (PLEG): container finished" podID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" containerID="050f6964e5cd767cf79a2ecd054c442d972a75a8c22c0e41a23da6992832e53e" exitCode=0 Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.821207 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z5p9" event={"ID":"ba2d8579-0ecb-46a2-950b-c48d660ed7ab","Type":"ContainerDied","Data":"050f6964e5cd767cf79a2ecd054c442d972a75a8c22c0e41a23da6992832e53e"} Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.828846 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvfqb" event={"ID":"2424e499-800f-481e-95e2-84257025d221","Type":"ContainerDied","Data":"2d43822aea2c9322a313dc58d7efd58ca721cb31d3b22fc37e09a961af79c440"} Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.828929 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wvfqb" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.829470 4945 scope.go:117] "RemoveContainer" containerID="c4d0ef4527004805103029e330db7e7010e0782dd9ee2c747cd75fea11102723" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.831316 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hrlc6"] Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.834406 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hrlc6"] Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.843302 4945 scope.go:117] "RemoveContainer" containerID="8d76a62a998efc5c5a21bdc496893e1bed4271e56d9e3d35dbbf6d57714e2d46" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.855830 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wvfqb"] Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.858656 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wvfqb"] Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.880711 4945 scope.go:117] "RemoveContainer" containerID="bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b" Oct 14 15:18:34 crc kubenswrapper[4945]: E1014 15:18:34.881137 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b\": container with ID starting with bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b not found: ID does not exist" containerID="bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.881180 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b"} err="failed to get container status \"bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b\": rpc error: code = NotFound desc = could not find container \"bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b\": container with ID starting with bdde1f9dd0dcc6fa5c2926d6f3686230be187cbdd73c2813dff8f458ee884e8b not found: ID does not exist" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.881216 4945 scope.go:117] "RemoveContainer" containerID="c4d0ef4527004805103029e330db7e7010e0782dd9ee2c747cd75fea11102723" Oct 14 15:18:34 crc kubenswrapper[4945]: E1014 15:18:34.881545 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4d0ef4527004805103029e330db7e7010e0782dd9ee2c747cd75fea11102723\": container with ID starting with c4d0ef4527004805103029e330db7e7010e0782dd9ee2c747cd75fea11102723 not found: ID does not exist" containerID="c4d0ef4527004805103029e330db7e7010e0782dd9ee2c747cd75fea11102723" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.881576 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4d0ef4527004805103029e330db7e7010e0782dd9ee2c747cd75fea11102723"} err="failed to get container status \"c4d0ef4527004805103029e330db7e7010e0782dd9ee2c747cd75fea11102723\": rpc error: code = NotFound desc = could not find container \"c4d0ef4527004805103029e330db7e7010e0782dd9ee2c747cd75fea11102723\": container with ID starting with c4d0ef4527004805103029e330db7e7010e0782dd9ee2c747cd75fea11102723 not found: ID does not exist" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.881602 4945 scope.go:117] "RemoveContainer" containerID="8d76a62a998efc5c5a21bdc496893e1bed4271e56d9e3d35dbbf6d57714e2d46" Oct 14 15:18:34 crc kubenswrapper[4945]: E1014 15:18:34.881940 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d76a62a998efc5c5a21bdc496893e1bed4271e56d9e3d35dbbf6d57714e2d46\": container with ID starting with 8d76a62a998efc5c5a21bdc496893e1bed4271e56d9e3d35dbbf6d57714e2d46 not found: ID does not exist" containerID="8d76a62a998efc5c5a21bdc496893e1bed4271e56d9e3d35dbbf6d57714e2d46" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.881965 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d76a62a998efc5c5a21bdc496893e1bed4271e56d9e3d35dbbf6d57714e2d46"} err="failed to get container status \"8d76a62a998efc5c5a21bdc496893e1bed4271e56d9e3d35dbbf6d57714e2d46\": rpc error: code = NotFound desc = could not find container \"8d76a62a998efc5c5a21bdc496893e1bed4271e56d9e3d35dbbf6d57714e2d46\": container with ID starting with 8d76a62a998efc5c5a21bdc496893e1bed4271e56d9e3d35dbbf6d57714e2d46 not found: ID does not exist" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.881981 4945 scope.go:117] "RemoveContainer" containerID="e4f1e4f5e2dfa4ab90e73fd43ae6ddc896f631a210dc262517aa84beef44203a" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.894595 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2424e499-800f-481e-95e2-84257025d221-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.897547 4945 scope.go:117] "RemoveContainer" containerID="f4d5b80642a6cfbc6c43ebeb0909605ed632b15c25c51fea5b78ed506ab9cb71" Oct 14 15:18:34 crc kubenswrapper[4945]: I1014 15:18:34.910311 4945 scope.go:117] "RemoveContainer" containerID="2c7420b342a6bc99d2d76dd72f7ba10db3bf359f30009ab6835a99f8b3a82843" Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.233633 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.400035 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmnsl\" (UniqueName: \"kubernetes.io/projected/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-kube-api-access-tmnsl\") pod \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\" (UID: \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\") " Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.400120 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-utilities\") pod \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\" (UID: \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\") " Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.400170 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-catalog-content\") pod \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\" (UID: \"ba2d8579-0ecb-46a2-950b-c48d660ed7ab\") " Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.400861 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-utilities" (OuterVolumeSpecName: "utilities") pod "ba2d8579-0ecb-46a2-950b-c48d660ed7ab" (UID: "ba2d8579-0ecb-46a2-950b-c48d660ed7ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.406056 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-kube-api-access-tmnsl" (OuterVolumeSpecName: "kube-api-access-tmnsl") pod "ba2d8579-0ecb-46a2-950b-c48d660ed7ab" (UID: "ba2d8579-0ecb-46a2-950b-c48d660ed7ab"). InnerVolumeSpecName "kube-api-access-tmnsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.460410 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba2d8579-0ecb-46a2-950b-c48d660ed7ab" (UID: "ba2d8579-0ecb-46a2-950b-c48d660ed7ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.501725 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmnsl\" (UniqueName: \"kubernetes.io/projected/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-kube-api-access-tmnsl\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.501764 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.501777 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba2d8579-0ecb-46a2-950b-c48d660ed7ab-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.838783 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z5p9" event={"ID":"ba2d8579-0ecb-46a2-950b-c48d660ed7ab","Type":"ContainerDied","Data":"d11fd99ea853960d888477eda5794744f2c555fadc4a700a60bf908adf98ec0d"} Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.838828 4945 scope.go:117] "RemoveContainer" containerID="050f6964e5cd767cf79a2ecd054c442d972a75a8c22c0e41a23da6992832e53e" Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.838908 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z5p9" Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.851452 4945 scope.go:117] "RemoveContainer" containerID="df23c4e4922478bda19b51688e38fd7314803f26cae413bfbbd6ff0a33ad765f" Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.866130 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8z5p9"] Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.868460 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8z5p9"] Oct 14 15:18:35 crc kubenswrapper[4945]: I1014 15:18:35.879644 4945 scope.go:117] "RemoveContainer" containerID="1eb0e4e31d78e600e87e9d170c0db58f48e5f04c1291d986b32e5ea2ad0d324b" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.200819 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-prqws"] Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.201094 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-prqws" podUID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" containerName="registry-server" containerID="cri-o://e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3" gracePeriod=2 Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.599671 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.715280 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9gzn\" (UniqueName: \"kubernetes.io/projected/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-kube-api-access-j9gzn\") pod \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\" (UID: \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\") " Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.715339 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-catalog-content\") pod \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\" (UID: \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\") " Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.715403 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-utilities\") pod \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\" (UID: \"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8\") " Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.716173 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-utilities" (OuterVolumeSpecName: "utilities") pod "a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" (UID: "a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.721995 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-kube-api-access-j9gzn" (OuterVolumeSpecName: "kube-api-access-j9gzn") pod "a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" (UID: "a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8"). InnerVolumeSpecName "kube-api-access-j9gzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.729803 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" (UID: "a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.771410 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2424e499-800f-481e-95e2-84257025d221" path="/var/lib/kubelet/pods/2424e499-800f-481e-95e2-84257025d221/volumes" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.771989 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" path="/var/lib/kubelet/pods/40554eb6-5285-4ae4-8e2b-ad6f546c0b53/volumes" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.772516 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" path="/var/lib/kubelet/pods/ba2d8579-0ecb-46a2-950b-c48d660ed7ab/volumes" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.816860 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.816924 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.816938 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9gzn\" (UniqueName: \"kubernetes.io/projected/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8-kube-api-access-j9gzn\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.847915 4945 generic.go:334] "Generic (PLEG): container finished" podID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" containerID="e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3" exitCode=0 Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.847972 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-prqws" event={"ID":"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8","Type":"ContainerDied","Data":"e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3"} Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.848015 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-prqws" event={"ID":"a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8","Type":"ContainerDied","Data":"14994abe2490ef9a5faaf6d94e55d98a5c8ab89f3539b3f42d137bcd007b837b"} Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.848023 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-prqws" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.848036 4945 scope.go:117] "RemoveContainer" containerID="e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.870304 4945 scope.go:117] "RemoveContainer" containerID="98229a8761089893f1ece194ecb43cd2702c5209c221caadac1c55d1356c8704" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.871012 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-prqws"] Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.875515 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-prqws"] Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.882614 4945 scope.go:117] "RemoveContainer" containerID="5d97db5d81371a633e2a299581dadae6123f02edb083acbc9f91e5eb5270b394" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.894655 4945 scope.go:117] "RemoveContainer" containerID="e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3" Oct 14 15:18:36 crc kubenswrapper[4945]: E1014 15:18:36.895003 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3\": container with ID starting with e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3 not found: ID does not exist" containerID="e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.895041 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3"} err="failed to get container status \"e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3\": rpc error: code = NotFound desc = could not find container \"e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3\": container with ID starting with e4db2fe9600c985ee2170421e094c614c03a68d4abaa7c8b95ecdf8ff796d4b3 not found: ID does not exist" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.895069 4945 scope.go:117] "RemoveContainer" containerID="98229a8761089893f1ece194ecb43cd2702c5209c221caadac1c55d1356c8704" Oct 14 15:18:36 crc kubenswrapper[4945]: E1014 15:18:36.895582 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98229a8761089893f1ece194ecb43cd2702c5209c221caadac1c55d1356c8704\": container with ID starting with 98229a8761089893f1ece194ecb43cd2702c5209c221caadac1c55d1356c8704 not found: ID does not exist" containerID="98229a8761089893f1ece194ecb43cd2702c5209c221caadac1c55d1356c8704" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.895620 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98229a8761089893f1ece194ecb43cd2702c5209c221caadac1c55d1356c8704"} err="failed to get container status \"98229a8761089893f1ece194ecb43cd2702c5209c221caadac1c55d1356c8704\": rpc error: code = NotFound desc = could not find container \"98229a8761089893f1ece194ecb43cd2702c5209c221caadac1c55d1356c8704\": container with ID starting with 98229a8761089893f1ece194ecb43cd2702c5209c221caadac1c55d1356c8704 not found: ID does not exist" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.895637 4945 scope.go:117] "RemoveContainer" containerID="5d97db5d81371a633e2a299581dadae6123f02edb083acbc9f91e5eb5270b394" Oct 14 15:18:36 crc kubenswrapper[4945]: E1014 15:18:36.896156 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d97db5d81371a633e2a299581dadae6123f02edb083acbc9f91e5eb5270b394\": container with ID starting with 5d97db5d81371a633e2a299581dadae6123f02edb083acbc9f91e5eb5270b394 not found: ID does not exist" containerID="5d97db5d81371a633e2a299581dadae6123f02edb083acbc9f91e5eb5270b394" Oct 14 15:18:36 crc kubenswrapper[4945]: I1014 15:18:36.896203 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d97db5d81371a633e2a299581dadae6123f02edb083acbc9f91e5eb5270b394"} err="failed to get container status \"5d97db5d81371a633e2a299581dadae6123f02edb083acbc9f91e5eb5270b394\": rpc error: code = NotFound desc = could not find container \"5d97db5d81371a633e2a299581dadae6123f02edb083acbc9f91e5eb5270b394\": container with ID starting with 5d97db5d81371a633e2a299581dadae6123f02edb083acbc9f91e5eb5270b394 not found: ID does not exist" Oct 14 15:18:38 crc kubenswrapper[4945]: I1014 15:18:38.769915 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" path="/var/lib/kubelet/pods/a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8/volumes" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.562545 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mq2vp"] Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563231 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2424e499-800f-481e-95e2-84257025d221" containerName="extract-content" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563244 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2424e499-800f-481e-95e2-84257025d221" containerName="extract-content" Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563255 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2424e499-800f-481e-95e2-84257025d221" containerName="extract-utilities" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563260 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2424e499-800f-481e-95e2-84257025d221" containerName="extract-utilities" Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563267 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" containerName="extract-utilities" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563274 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" containerName="extract-utilities" Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563281 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc1d96aa-1a35-45ee-99f9-2f28440fa345" containerName="pruner" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563286 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc1d96aa-1a35-45ee-99f9-2f28440fa345" containerName="pruner" Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563294 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" containerName="extract-utilities" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563299 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" containerName="extract-utilities" Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563306 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2424e499-800f-481e-95e2-84257025d221" containerName="registry-server" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563312 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2424e499-800f-481e-95e2-84257025d221" containerName="registry-server" Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563319 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" containerName="extract-utilities" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563325 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" containerName="extract-utilities" Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563333 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" containerName="registry-server" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563338 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" containerName="registry-server" Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563348 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" containerName="registry-server" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563354 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" containerName="registry-server" Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563360 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" containerName="extract-content" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563366 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" containerName="extract-content" Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563373 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" containerName="extract-content" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563379 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" containerName="extract-content" Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563386 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" containerName="registry-server" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563392 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" containerName="registry-server" Oct 14 15:18:44 crc kubenswrapper[4945]: E1014 15:18:44.563403 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" containerName="extract-content" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563410 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" containerName="extract-content" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563489 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba2d8579-0ecb-46a2-950b-c48d660ed7ab" containerName="registry-server" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563497 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2424e499-800f-481e-95e2-84257025d221" containerName="registry-server" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563505 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a154a74e-14ee-43fd-a9f7-e2a8fbffdfe8" containerName="registry-server" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563515 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc1d96aa-1a35-45ee-99f9-2f28440fa345" containerName="pruner" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563522 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="40554eb6-5285-4ae4-8e2b-ad6f546c0b53" containerName="registry-server" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.563910 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.588591 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mq2vp"] Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.747330 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m95n\" (UniqueName: \"kubernetes.io/projected/653d58ea-157f-45a0-8c71-7640ff95a47b-kube-api-access-4m95n\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.747388 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.747453 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/653d58ea-157f-45a0-8c71-7640ff95a47b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.747483 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/653d58ea-157f-45a0-8c71-7640ff95a47b-registry-certificates\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.747528 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/653d58ea-157f-45a0-8c71-7640ff95a47b-trusted-ca\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.747554 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/653d58ea-157f-45a0-8c71-7640ff95a47b-bound-sa-token\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.747571 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/653d58ea-157f-45a0-8c71-7640ff95a47b-registry-tls\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.747647 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/653d58ea-157f-45a0-8c71-7640ff95a47b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.768838 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.849106 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/653d58ea-157f-45a0-8c71-7640ff95a47b-trusted-ca\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.849146 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/653d58ea-157f-45a0-8c71-7640ff95a47b-bound-sa-token\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.849165 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/653d58ea-157f-45a0-8c71-7640ff95a47b-registry-tls\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.849201 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/653d58ea-157f-45a0-8c71-7640ff95a47b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.849257 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m95n\" (UniqueName: \"kubernetes.io/projected/653d58ea-157f-45a0-8c71-7640ff95a47b-kube-api-access-4m95n\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.849287 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/653d58ea-157f-45a0-8c71-7640ff95a47b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.849304 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/653d58ea-157f-45a0-8c71-7640ff95a47b-registry-certificates\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.850672 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/653d58ea-157f-45a0-8c71-7640ff95a47b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.850808 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/653d58ea-157f-45a0-8c71-7640ff95a47b-trusted-ca\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.851914 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/653d58ea-157f-45a0-8c71-7640ff95a47b-registry-certificates\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.856591 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/653d58ea-157f-45a0-8c71-7640ff95a47b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.856626 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/653d58ea-157f-45a0-8c71-7640ff95a47b-registry-tls\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.865340 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/653d58ea-157f-45a0-8c71-7640ff95a47b-bound-sa-token\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.868033 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m95n\" (UniqueName: \"kubernetes.io/projected/653d58ea-157f-45a0-8c71-7640ff95a47b-kube-api-access-4m95n\") pod \"image-registry-66df7c8f76-mq2vp\" (UID: \"653d58ea-157f-45a0-8c71-7640ff95a47b\") " pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:44 crc kubenswrapper[4945]: I1014 15:18:44.879202 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:45 crc kubenswrapper[4945]: I1014 15:18:45.303053 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mq2vp"] Oct 14 15:18:45 crc kubenswrapper[4945]: I1014 15:18:45.888442 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" event={"ID":"653d58ea-157f-45a0-8c71-7640ff95a47b","Type":"ContainerStarted","Data":"b360b45768e3589fd9e08fc8f7be5b01eac7f883c95f611878f9cd0c940c9d93"} Oct 14 15:18:45 crc kubenswrapper[4945]: I1014 15:18:45.888804 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" event={"ID":"653d58ea-157f-45a0-8c71-7640ff95a47b","Type":"ContainerStarted","Data":"fa8b5df8c5eec7cd501fa0edf896f83c30eb58449d4ebc2fbde6e9e3f9bf2ef4"} Oct 14 15:18:45 crc kubenswrapper[4945]: I1014 15:18:45.888824 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:18:45 crc kubenswrapper[4945]: I1014 15:18:45.905067 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" podStartSLOduration=1.9050508480000001 podStartE2EDuration="1.905050848s" podCreationTimestamp="2025-10-14 15:18:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:18:45.903954653 +0000 UTC m=+255.888003021" watchObservedRunningTime="2025-10-14 15:18:45.905050848 +0000 UTC m=+255.889099216" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.505622 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tfj4f"] Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.506453 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tfj4f" podUID="ec783233-16ca-4e17-8923-71455b2056d7" containerName="registry-server" containerID="cri-o://2c274dd86cc88755f803c9de36dbd4388371808e03e060a0d26f6af565114ccf" gracePeriod=30 Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.514511 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vs4wr"] Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.514758 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vs4wr" podUID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" containerName="registry-server" containerID="cri-o://97e4d44348b3e6dea11ce84bee2e6afadb6b6534f3e24235f30d7c34fc04f864" gracePeriod=30 Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.527887 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2kwtj"] Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.528131 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" podUID="98994fb5-4f28-4381-9ae2-4c734bf79c5c" containerName="marketplace-operator" containerID="cri-o://325e630298c84c6bd6ad7affff339d43dc7e628aa02b16abf28c73bdbc17bb9d" gracePeriod=30 Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.548234 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b27lt"] Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.548569 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b27lt" podUID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" containerName="registry-server" containerID="cri-o://2278bb953c1e948d862bb06a80c4c88b3b14f37b9785912dfea06a88d9eb31d9" gracePeriod=30 Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.549063 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5lphm"] Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.549212 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5lphm" podUID="73627872-659a-47d9-8aa5-c55682facdc7" containerName="registry-server" containerID="cri-o://1411170f7c2bd1ae8ec61f07c42baa61524c6affad3a7ec464f6240554d8edc7" gracePeriod=30 Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.553988 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hdrgb"] Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.554770 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.556690 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hdrgb"] Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.710816 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8ffh\" (UniqueName: \"kubernetes.io/projected/1139d59d-b224-467d-849e-af47cba636e6-kube-api-access-s8ffh\") pod \"marketplace-operator-79b997595-hdrgb\" (UID: \"1139d59d-b224-467d-849e-af47cba636e6\") " pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.711188 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1139d59d-b224-467d-849e-af47cba636e6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hdrgb\" (UID: \"1139d59d-b224-467d-849e-af47cba636e6\") " pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.711258 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1139d59d-b224-467d-849e-af47cba636e6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hdrgb\" (UID: \"1139d59d-b224-467d-849e-af47cba636e6\") " pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.812511 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1139d59d-b224-467d-849e-af47cba636e6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hdrgb\" (UID: \"1139d59d-b224-467d-849e-af47cba636e6\") " pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.812563 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1139d59d-b224-467d-849e-af47cba636e6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hdrgb\" (UID: \"1139d59d-b224-467d-849e-af47cba636e6\") " pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.812604 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8ffh\" (UniqueName: \"kubernetes.io/projected/1139d59d-b224-467d-849e-af47cba636e6-kube-api-access-s8ffh\") pod \"marketplace-operator-79b997595-hdrgb\" (UID: \"1139d59d-b224-467d-849e-af47cba636e6\") " pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.814458 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1139d59d-b224-467d-849e-af47cba636e6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hdrgb\" (UID: \"1139d59d-b224-467d-849e-af47cba636e6\") " pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.818211 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1139d59d-b224-467d-849e-af47cba636e6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hdrgb\" (UID: \"1139d59d-b224-467d-849e-af47cba636e6\") " pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.826171 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8ffh\" (UniqueName: \"kubernetes.io/projected/1139d59d-b224-467d-849e-af47cba636e6-kube-api-access-s8ffh\") pod \"marketplace-operator-79b997595-hdrgb\" (UID: \"1139d59d-b224-467d-849e-af47cba636e6\") " pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.872986 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.915395 4945 generic.go:334] "Generic (PLEG): container finished" podID="ec783233-16ca-4e17-8923-71455b2056d7" containerID="2c274dd86cc88755f803c9de36dbd4388371808e03e060a0d26f6af565114ccf" exitCode=0 Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.915512 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfj4f" event={"ID":"ec783233-16ca-4e17-8923-71455b2056d7","Type":"ContainerDied","Data":"2c274dd86cc88755f803c9de36dbd4388371808e03e060a0d26f6af565114ccf"} Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.920804 4945 generic.go:334] "Generic (PLEG): container finished" podID="73627872-659a-47d9-8aa5-c55682facdc7" containerID="1411170f7c2bd1ae8ec61f07c42baa61524c6affad3a7ec464f6240554d8edc7" exitCode=0 Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.920845 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lphm" event={"ID":"73627872-659a-47d9-8aa5-c55682facdc7","Type":"ContainerDied","Data":"1411170f7c2bd1ae8ec61f07c42baa61524c6affad3a7ec464f6240554d8edc7"} Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.924370 4945 generic.go:334] "Generic (PLEG): container finished" podID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" containerID="97e4d44348b3e6dea11ce84bee2e6afadb6b6534f3e24235f30d7c34fc04f864" exitCode=0 Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.924414 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vs4wr" event={"ID":"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039","Type":"ContainerDied","Data":"97e4d44348b3e6dea11ce84bee2e6afadb6b6534f3e24235f30d7c34fc04f864"} Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.924433 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vs4wr" event={"ID":"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039","Type":"ContainerDied","Data":"0a555505cbef0d00e1a2d3f9639cb30f5473cd21b8b987669ddf5f278d3a5c50"} Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.924448 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a555505cbef0d00e1a2d3f9639cb30f5473cd21b8b987669ddf5f278d3a5c50" Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.929828 4945 generic.go:334] "Generic (PLEG): container finished" podID="98994fb5-4f28-4381-9ae2-4c734bf79c5c" containerID="325e630298c84c6bd6ad7affff339d43dc7e628aa02b16abf28c73bdbc17bb9d" exitCode=0 Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.929904 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" event={"ID":"98994fb5-4f28-4381-9ae2-4c734bf79c5c","Type":"ContainerDied","Data":"325e630298c84c6bd6ad7affff339d43dc7e628aa02b16abf28c73bdbc17bb9d"} Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.935753 4945 generic.go:334] "Generic (PLEG): container finished" podID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" containerID="2278bb953c1e948d862bb06a80c4c88b3b14f37b9785912dfea06a88d9eb31d9" exitCode=0 Oct 14 15:18:49 crc kubenswrapper[4945]: I1014 15:18:49.935775 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b27lt" event={"ID":"27eaa4e2-7ffd-41b3-9d3d-418b43669de8","Type":"ContainerDied","Data":"2278bb953c1e948d862bb06a80c4c88b3b14f37b9785912dfea06a88d9eb31d9"} Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.004201 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.026539 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.034788 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.049465 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.074859 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.116584 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec783233-16ca-4e17-8923-71455b2056d7-utilities\") pod \"ec783233-16ca-4e17-8923-71455b2056d7\" (UID: \"ec783233-16ca-4e17-8923-71455b2056d7\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.116643 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnpc4\" (UniqueName: \"kubernetes.io/projected/98994fb5-4f28-4381-9ae2-4c734bf79c5c-kube-api-access-rnpc4\") pod \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\" (UID: \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.116671 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/98994fb5-4f28-4381-9ae2-4c734bf79c5c-marketplace-operator-metrics\") pod \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\" (UID: \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.116717 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-utilities\") pod \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\" (UID: \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.116732 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-catalog-content\") pod \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\" (UID: \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.116747 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkwqk\" (UniqueName: \"kubernetes.io/projected/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-kube-api-access-lkwqk\") pod \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\" (UID: \"5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.116769 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-utilities\") pod \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\" (UID: \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.116787 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec783233-16ca-4e17-8923-71455b2056d7-catalog-content\") pod \"ec783233-16ca-4e17-8923-71455b2056d7\" (UID: \"ec783233-16ca-4e17-8923-71455b2056d7\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.116829 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mm4p\" (UniqueName: \"kubernetes.io/projected/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-kube-api-access-4mm4p\") pod \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\" (UID: \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.116845 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-catalog-content\") pod \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\" (UID: \"27eaa4e2-7ffd-41b3-9d3d-418b43669de8\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.116869 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98994fb5-4f28-4381-9ae2-4c734bf79c5c-marketplace-trusted-ca\") pod \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\" (UID: \"98994fb5-4f28-4381-9ae2-4c734bf79c5c\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.116904 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klwwd\" (UniqueName: \"kubernetes.io/projected/ec783233-16ca-4e17-8923-71455b2056d7-kube-api-access-klwwd\") pod \"ec783233-16ca-4e17-8923-71455b2056d7\" (UID: \"ec783233-16ca-4e17-8923-71455b2056d7\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.117676 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec783233-16ca-4e17-8923-71455b2056d7-utilities" (OuterVolumeSpecName: "utilities") pod "ec783233-16ca-4e17-8923-71455b2056d7" (UID: "ec783233-16ca-4e17-8923-71455b2056d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.118477 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-utilities" (OuterVolumeSpecName: "utilities") pod "5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" (UID: "5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.118914 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98994fb5-4f28-4381-9ae2-4c734bf79c5c-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "98994fb5-4f28-4381-9ae2-4c734bf79c5c" (UID: "98994fb5-4f28-4381-9ae2-4c734bf79c5c"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.119375 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-utilities" (OuterVolumeSpecName: "utilities") pod "27eaa4e2-7ffd-41b3-9d3d-418b43669de8" (UID: "27eaa4e2-7ffd-41b3-9d3d-418b43669de8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.122368 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec783233-16ca-4e17-8923-71455b2056d7-kube-api-access-klwwd" (OuterVolumeSpecName: "kube-api-access-klwwd") pod "ec783233-16ca-4e17-8923-71455b2056d7" (UID: "ec783233-16ca-4e17-8923-71455b2056d7"). InnerVolumeSpecName "kube-api-access-klwwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.122817 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98994fb5-4f28-4381-9ae2-4c734bf79c5c-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "98994fb5-4f28-4381-9ae2-4c734bf79c5c" (UID: "98994fb5-4f28-4381-9ae2-4c734bf79c5c"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.123072 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-kube-api-access-lkwqk" (OuterVolumeSpecName: "kube-api-access-lkwqk") pod "5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" (UID: "5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039"). InnerVolumeSpecName "kube-api-access-lkwqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.123229 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-kube-api-access-4mm4p" (OuterVolumeSpecName: "kube-api-access-4mm4p") pod "27eaa4e2-7ffd-41b3-9d3d-418b43669de8" (UID: "27eaa4e2-7ffd-41b3-9d3d-418b43669de8"). InnerVolumeSpecName "kube-api-access-4mm4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.124064 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98994fb5-4f28-4381-9ae2-4c734bf79c5c-kube-api-access-rnpc4" (OuterVolumeSpecName: "kube-api-access-rnpc4") pod "98994fb5-4f28-4381-9ae2-4c734bf79c5c" (UID: "98994fb5-4f28-4381-9ae2-4c734bf79c5c"). InnerVolumeSpecName "kube-api-access-rnpc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.132869 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27eaa4e2-7ffd-41b3-9d3d-418b43669de8" (UID: "27eaa4e2-7ffd-41b3-9d3d-418b43669de8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.177994 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" (UID: "5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.186159 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec783233-16ca-4e17-8923-71455b2056d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec783233-16ca-4e17-8923-71455b2056d7" (UID: "ec783233-16ca-4e17-8923-71455b2056d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218464 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd5kh\" (UniqueName: \"kubernetes.io/projected/73627872-659a-47d9-8aa5-c55682facdc7-kube-api-access-fd5kh\") pod \"73627872-659a-47d9-8aa5-c55682facdc7\" (UID: \"73627872-659a-47d9-8aa5-c55682facdc7\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218599 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73627872-659a-47d9-8aa5-c55682facdc7-catalog-content\") pod \"73627872-659a-47d9-8aa5-c55682facdc7\" (UID: \"73627872-659a-47d9-8aa5-c55682facdc7\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218633 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73627872-659a-47d9-8aa5-c55682facdc7-utilities\") pod \"73627872-659a-47d9-8aa5-c55682facdc7\" (UID: \"73627872-659a-47d9-8aa5-c55682facdc7\") " Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218846 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218864 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkwqk\" (UniqueName: \"kubernetes.io/projected/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-kube-api-access-lkwqk\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218901 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218912 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218920 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec783233-16ca-4e17-8923-71455b2056d7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218929 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mm4p\" (UniqueName: \"kubernetes.io/projected/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-kube-api-access-4mm4p\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218937 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27eaa4e2-7ffd-41b3-9d3d-418b43669de8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218947 4945 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98994fb5-4f28-4381-9ae2-4c734bf79c5c-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218955 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klwwd\" (UniqueName: \"kubernetes.io/projected/ec783233-16ca-4e17-8923-71455b2056d7-kube-api-access-klwwd\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218962 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec783233-16ca-4e17-8923-71455b2056d7-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218971 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnpc4\" (UniqueName: \"kubernetes.io/projected/98994fb5-4f28-4381-9ae2-4c734bf79c5c-kube-api-access-rnpc4\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.218979 4945 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/98994fb5-4f28-4381-9ae2-4c734bf79c5c-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.219385 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73627872-659a-47d9-8aa5-c55682facdc7-utilities" (OuterVolumeSpecName: "utilities") pod "73627872-659a-47d9-8aa5-c55682facdc7" (UID: "73627872-659a-47d9-8aa5-c55682facdc7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.221414 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73627872-659a-47d9-8aa5-c55682facdc7-kube-api-access-fd5kh" (OuterVolumeSpecName: "kube-api-access-fd5kh") pod "73627872-659a-47d9-8aa5-c55682facdc7" (UID: "73627872-659a-47d9-8aa5-c55682facdc7"). InnerVolumeSpecName "kube-api-access-fd5kh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.315680 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73627872-659a-47d9-8aa5-c55682facdc7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73627872-659a-47d9-8aa5-c55682facdc7" (UID: "73627872-659a-47d9-8aa5-c55682facdc7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.318607 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hdrgb"] Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.320321 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73627872-659a-47d9-8aa5-c55682facdc7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.320418 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73627872-659a-47d9-8aa5-c55682facdc7-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.320490 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd5kh\" (UniqueName: \"kubernetes.io/projected/73627872-659a-47d9-8aa5-c55682facdc7-kube-api-access-fd5kh\") on node \"crc\" DevicePath \"\"" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.481189 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9zzcq"] Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.944811 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lphm" event={"ID":"73627872-659a-47d9-8aa5-c55682facdc7","Type":"ContainerDied","Data":"fec1210ef8844855ea194102d51806a3dfb866539957ba3ef118f41202efb2ad"} Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.945250 4945 scope.go:117] "RemoveContainer" containerID="1411170f7c2bd1ae8ec61f07c42baa61524c6affad3a7ec464f6240554d8edc7" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.944845 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5lphm" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.946786 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" event={"ID":"1139d59d-b224-467d-849e-af47cba636e6","Type":"ContainerStarted","Data":"4f2e92e56107667b321aa0a8d51443571834b69254b2abb716c4ad4633343252"} Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.946836 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" event={"ID":"1139d59d-b224-467d-849e-af47cba636e6","Type":"ContainerStarted","Data":"6de999c4c2b52f2892c08b0c4093aa05c2c482c8ed2b574df72248ed6fed5b87"} Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.947278 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.948437 4945 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-hdrgb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.948482 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" podUID="1139d59d-b224-467d-849e-af47cba636e6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.949395 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.949409 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2kwtj" event={"ID":"98994fb5-4f28-4381-9ae2-4c734bf79c5c","Type":"ContainerDied","Data":"03ed83a4e5fbb9e657f904694242a6c57c79a96d47b4a3094938e93242f5bbf1"} Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.952791 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b27lt" event={"ID":"27eaa4e2-7ffd-41b3-9d3d-418b43669de8","Type":"ContainerDied","Data":"6384aaee345c53151ec52a0cdcab05d97e9e340f03023a12f14c7effa334b6fa"} Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.952938 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b27lt" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.955808 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tfj4f" event={"ID":"ec783233-16ca-4e17-8923-71455b2056d7","Type":"ContainerDied","Data":"0232d42e0d70e53fbbc53df8ffceb40119496e95f239e46502987f7c682c35eb"} Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.955823 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vs4wr" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.955815 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tfj4f" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.970234 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" podStartSLOduration=1.970213578 podStartE2EDuration="1.970213578s" podCreationTimestamp="2025-10-14 15:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:18:50.968951768 +0000 UTC m=+260.953000136" watchObservedRunningTime="2025-10-14 15:18:50.970213578 +0000 UTC m=+260.954261946" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.973383 4945 scope.go:117] "RemoveContainer" containerID="38084c83fb9d2954669b62c23f53d4ac610dcae410397f17a35896126405f0b9" Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.983028 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5lphm"] Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.986573 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5lphm"] Oct 14 15:18:50 crc kubenswrapper[4945]: I1014 15:18:50.998107 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2kwtj"] Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.001297 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2kwtj"] Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.003272 4945 scope.go:117] "RemoveContainer" containerID="e19e9407cbcebb564ec22acbbc1d04496e37b5e45fd9fe7cc940c39381292cd5" Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.005759 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b27lt"] Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.008697 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b27lt"] Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.015487 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vs4wr"] Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.018997 4945 scope.go:117] "RemoveContainer" containerID="325e630298c84c6bd6ad7affff339d43dc7e628aa02b16abf28c73bdbc17bb9d" Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.020233 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vs4wr"] Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.036894 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tfj4f"] Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.044853 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tfj4f"] Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.046774 4945 scope.go:117] "RemoveContainer" containerID="2278bb953c1e948d862bb06a80c4c88b3b14f37b9785912dfea06a88d9eb31d9" Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.064211 4945 scope.go:117] "RemoveContainer" containerID="c73db62ba2cebd6ed88186390619656b58ebbffbc1154ab7a657d3b591f51a23" Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.079274 4945 scope.go:117] "RemoveContainer" containerID="20c90023e137420bf01e365fc2c7cfb6d03619a7f5f3463dcae74761b09a6175" Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.097619 4945 scope.go:117] "RemoveContainer" containerID="2c274dd86cc88755f803c9de36dbd4388371808e03e060a0d26f6af565114ccf" Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.114853 4945 scope.go:117] "RemoveContainer" containerID="7bf4a9f23db2d0f71bc5cd308b83d01cb61b4d2a57986c153bc5f6e1b67bc851" Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.129244 4945 scope.go:117] "RemoveContainer" containerID="5ec5f64ebde6d98946d7c2a23464c2e4bc65b08e560315ac1ba8cd98f71a2600" Oct 14 15:18:51 crc kubenswrapper[4945]: I1014 15:18:51.970855 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hdrgb" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.314624 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m9lt9"] Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315187 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" containerName="registry-server" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315204 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" containerName="registry-server" Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315215 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" containerName="extract-content" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315223 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" containerName="extract-content" Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315236 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" containerName="extract-utilities" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315245 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" containerName="extract-utilities" Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315257 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73627872-659a-47d9-8aa5-c55682facdc7" containerName="extract-utilities" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315264 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="73627872-659a-47d9-8aa5-c55682facdc7" containerName="extract-utilities" Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315273 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73627872-659a-47d9-8aa5-c55682facdc7" containerName="extract-content" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315280 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="73627872-659a-47d9-8aa5-c55682facdc7" containerName="extract-content" Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315289 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" containerName="extract-content" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315296 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" containerName="extract-content" Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315307 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec783233-16ca-4e17-8923-71455b2056d7" containerName="registry-server" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315315 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec783233-16ca-4e17-8923-71455b2056d7" containerName="registry-server" Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315322 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec783233-16ca-4e17-8923-71455b2056d7" containerName="extract-utilities" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315330 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec783233-16ca-4e17-8923-71455b2056d7" containerName="extract-utilities" Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315341 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98994fb5-4f28-4381-9ae2-4c734bf79c5c" containerName="marketplace-operator" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315349 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="98994fb5-4f28-4381-9ae2-4c734bf79c5c" containerName="marketplace-operator" Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315363 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" containerName="extract-utilities" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315370 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" containerName="extract-utilities" Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315384 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec783233-16ca-4e17-8923-71455b2056d7" containerName="extract-content" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315392 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec783233-16ca-4e17-8923-71455b2056d7" containerName="extract-content" Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315400 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73627872-659a-47d9-8aa5-c55682facdc7" containerName="registry-server" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315408 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="73627872-659a-47d9-8aa5-c55682facdc7" containerName="registry-server" Oct 14 15:18:52 crc kubenswrapper[4945]: E1014 15:18:52.315418 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" containerName="registry-server" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315426 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" containerName="registry-server" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315562 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec783233-16ca-4e17-8923-71455b2056d7" containerName="registry-server" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315580 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="98994fb5-4f28-4381-9ae2-4c734bf79c5c" containerName="marketplace-operator" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315590 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="73627872-659a-47d9-8aa5-c55682facdc7" containerName="registry-server" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315599 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" containerName="registry-server" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.315609 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" containerName="registry-server" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.316439 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.319737 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.323294 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m9lt9"] Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.450571 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f2803bc-a79f-4af4-99c7-f372b3b07bc8-catalog-content\") pod \"redhat-marketplace-m9lt9\" (UID: \"0f2803bc-a79f-4af4-99c7-f372b3b07bc8\") " pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.450617 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f2803bc-a79f-4af4-99c7-f372b3b07bc8-utilities\") pod \"redhat-marketplace-m9lt9\" (UID: \"0f2803bc-a79f-4af4-99c7-f372b3b07bc8\") " pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.450703 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhp9c\" (UniqueName: \"kubernetes.io/projected/0f2803bc-a79f-4af4-99c7-f372b3b07bc8-kube-api-access-xhp9c\") pod \"redhat-marketplace-m9lt9\" (UID: \"0f2803bc-a79f-4af4-99c7-f372b3b07bc8\") " pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.552342 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhp9c\" (UniqueName: \"kubernetes.io/projected/0f2803bc-a79f-4af4-99c7-f372b3b07bc8-kube-api-access-xhp9c\") pod \"redhat-marketplace-m9lt9\" (UID: \"0f2803bc-a79f-4af4-99c7-f372b3b07bc8\") " pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.552435 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f2803bc-a79f-4af4-99c7-f372b3b07bc8-catalog-content\") pod \"redhat-marketplace-m9lt9\" (UID: \"0f2803bc-a79f-4af4-99c7-f372b3b07bc8\") " pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.552470 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f2803bc-a79f-4af4-99c7-f372b3b07bc8-utilities\") pod \"redhat-marketplace-m9lt9\" (UID: \"0f2803bc-a79f-4af4-99c7-f372b3b07bc8\") " pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.552806 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f2803bc-a79f-4af4-99c7-f372b3b07bc8-catalog-content\") pod \"redhat-marketplace-m9lt9\" (UID: \"0f2803bc-a79f-4af4-99c7-f372b3b07bc8\") " pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.552889 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f2803bc-a79f-4af4-99c7-f372b3b07bc8-utilities\") pod \"redhat-marketplace-m9lt9\" (UID: \"0f2803bc-a79f-4af4-99c7-f372b3b07bc8\") " pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.574062 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhp9c\" (UniqueName: \"kubernetes.io/projected/0f2803bc-a79f-4af4-99c7-f372b3b07bc8-kube-api-access-xhp9c\") pod \"redhat-marketplace-m9lt9\" (UID: \"0f2803bc-a79f-4af4-99c7-f372b3b07bc8\") " pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.637151 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.769983 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27eaa4e2-7ffd-41b3-9d3d-418b43669de8" path="/var/lib/kubelet/pods/27eaa4e2-7ffd-41b3-9d3d-418b43669de8/volumes" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.771159 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039" path="/var/lib/kubelet/pods/5aaf49f3-2af5-4b0a-9ec7-f9ca24c27039/volumes" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.771717 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73627872-659a-47d9-8aa5-c55682facdc7" path="/var/lib/kubelet/pods/73627872-659a-47d9-8aa5-c55682facdc7/volumes" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.773157 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98994fb5-4f28-4381-9ae2-4c734bf79c5c" path="/var/lib/kubelet/pods/98994fb5-4f28-4381-9ae2-4c734bf79c5c/volumes" Oct 14 15:18:52 crc kubenswrapper[4945]: I1014 15:18:52.773782 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec783233-16ca-4e17-8923-71455b2056d7" path="/var/lib/kubelet/pods/ec783233-16ca-4e17-8923-71455b2056d7/volumes" Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.027344 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m9lt9"] Oct 14 15:18:53 crc kubenswrapper[4945]: W1014 15:18:53.031416 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f2803bc_a79f_4af4_99c7_f372b3b07bc8.slice/crio-9da47afff4a4066da1db699ce76a5117bcb6bfd693c518ad99018eb289a1760c WatchSource:0}: Error finding container 9da47afff4a4066da1db699ce76a5117bcb6bfd693c518ad99018eb289a1760c: Status 404 returned error can't find the container with id 9da47afff4a4066da1db699ce76a5117bcb6bfd693c518ad99018eb289a1760c Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.320517 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s9z56"] Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.322316 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.326824 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.332319 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s9z56"] Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.463931 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33acd0ef-71bb-467c-be47-0c93a694c8f2-catalog-content\") pod \"redhat-operators-s9z56\" (UID: \"33acd0ef-71bb-467c-be47-0c93a694c8f2\") " pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.464010 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33acd0ef-71bb-467c-be47-0c93a694c8f2-utilities\") pod \"redhat-operators-s9z56\" (UID: \"33acd0ef-71bb-467c-be47-0c93a694c8f2\") " pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.464170 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mnwl\" (UniqueName: \"kubernetes.io/projected/33acd0ef-71bb-467c-be47-0c93a694c8f2-kube-api-access-2mnwl\") pod \"redhat-operators-s9z56\" (UID: \"33acd0ef-71bb-467c-be47-0c93a694c8f2\") " pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.565223 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mnwl\" (UniqueName: \"kubernetes.io/projected/33acd0ef-71bb-467c-be47-0c93a694c8f2-kube-api-access-2mnwl\") pod \"redhat-operators-s9z56\" (UID: \"33acd0ef-71bb-467c-be47-0c93a694c8f2\") " pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.565302 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33acd0ef-71bb-467c-be47-0c93a694c8f2-catalog-content\") pod \"redhat-operators-s9z56\" (UID: \"33acd0ef-71bb-467c-be47-0c93a694c8f2\") " pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.565328 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33acd0ef-71bb-467c-be47-0c93a694c8f2-utilities\") pod \"redhat-operators-s9z56\" (UID: \"33acd0ef-71bb-467c-be47-0c93a694c8f2\") " pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.565682 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33acd0ef-71bb-467c-be47-0c93a694c8f2-utilities\") pod \"redhat-operators-s9z56\" (UID: \"33acd0ef-71bb-467c-be47-0c93a694c8f2\") " pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.565758 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33acd0ef-71bb-467c-be47-0c93a694c8f2-catalog-content\") pod \"redhat-operators-s9z56\" (UID: \"33acd0ef-71bb-467c-be47-0c93a694c8f2\") " pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.596562 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mnwl\" (UniqueName: \"kubernetes.io/projected/33acd0ef-71bb-467c-be47-0c93a694c8f2-kube-api-access-2mnwl\") pod \"redhat-operators-s9z56\" (UID: \"33acd0ef-71bb-467c-be47-0c93a694c8f2\") " pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:18:53 crc kubenswrapper[4945]: I1014 15:18:53.651116 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.015965 4945 generic.go:334] "Generic (PLEG): container finished" podID="0f2803bc-a79f-4af4-99c7-f372b3b07bc8" containerID="9f6e096a9ca2461a0198e6fb2319449c2b96cb78131f6a36a7e519911915a752" exitCode=0 Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.016210 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9lt9" event={"ID":"0f2803bc-a79f-4af4-99c7-f372b3b07bc8","Type":"ContainerDied","Data":"9f6e096a9ca2461a0198e6fb2319449c2b96cb78131f6a36a7e519911915a752"} Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.016236 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9lt9" event={"ID":"0f2803bc-a79f-4af4-99c7-f372b3b07bc8","Type":"ContainerStarted","Data":"9da47afff4a4066da1db699ce76a5117bcb6bfd693c518ad99018eb289a1760c"} Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.068822 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s9z56"] Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.717505 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v4cn2"] Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.718930 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.721349 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.728055 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v4cn2"] Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.821947 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63cb59cc-d19f-41e4-a839-1b67f86a5ffe-utilities\") pod \"community-operators-v4cn2\" (UID: \"63cb59cc-d19f-41e4-a839-1b67f86a5ffe\") " pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.822014 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh576\" (UniqueName: \"kubernetes.io/projected/63cb59cc-d19f-41e4-a839-1b67f86a5ffe-kube-api-access-vh576\") pod \"community-operators-v4cn2\" (UID: \"63cb59cc-d19f-41e4-a839-1b67f86a5ffe\") " pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.822237 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63cb59cc-d19f-41e4-a839-1b67f86a5ffe-catalog-content\") pod \"community-operators-v4cn2\" (UID: \"63cb59cc-d19f-41e4-a839-1b67f86a5ffe\") " pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.923072 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63cb59cc-d19f-41e4-a839-1b67f86a5ffe-utilities\") pod \"community-operators-v4cn2\" (UID: \"63cb59cc-d19f-41e4-a839-1b67f86a5ffe\") " pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.923131 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh576\" (UniqueName: \"kubernetes.io/projected/63cb59cc-d19f-41e4-a839-1b67f86a5ffe-kube-api-access-vh576\") pod \"community-operators-v4cn2\" (UID: \"63cb59cc-d19f-41e4-a839-1b67f86a5ffe\") " pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.923242 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63cb59cc-d19f-41e4-a839-1b67f86a5ffe-catalog-content\") pod \"community-operators-v4cn2\" (UID: \"63cb59cc-d19f-41e4-a839-1b67f86a5ffe\") " pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.923506 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63cb59cc-d19f-41e4-a839-1b67f86a5ffe-utilities\") pod \"community-operators-v4cn2\" (UID: \"63cb59cc-d19f-41e4-a839-1b67f86a5ffe\") " pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.923538 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63cb59cc-d19f-41e4-a839-1b67f86a5ffe-catalog-content\") pod \"community-operators-v4cn2\" (UID: \"63cb59cc-d19f-41e4-a839-1b67f86a5ffe\") " pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:18:54 crc kubenswrapper[4945]: I1014 15:18:54.941050 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh576\" (UniqueName: \"kubernetes.io/projected/63cb59cc-d19f-41e4-a839-1b67f86a5ffe-kube-api-access-vh576\") pod \"community-operators-v4cn2\" (UID: \"63cb59cc-d19f-41e4-a839-1b67f86a5ffe\") " pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.023506 4945 generic.go:334] "Generic (PLEG): container finished" podID="33acd0ef-71bb-467c-be47-0c93a694c8f2" containerID="02fec9c00230c6c0fcb51f6b8181fd943598de3ec7c468d25011d45a139b200d" exitCode=0 Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.023596 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9z56" event={"ID":"33acd0ef-71bb-467c-be47-0c93a694c8f2","Type":"ContainerDied","Data":"02fec9c00230c6c0fcb51f6b8181fd943598de3ec7c468d25011d45a139b200d"} Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.023636 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9z56" event={"ID":"33acd0ef-71bb-467c-be47-0c93a694c8f2","Type":"ContainerStarted","Data":"b758469bc62237c2f436a7c040785271c115636834399d3436ea3e40e80c3527"} Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.025842 4945 generic.go:334] "Generic (PLEG): container finished" podID="0f2803bc-a79f-4af4-99c7-f372b3b07bc8" containerID="0ed9a2244e21941824e7de899fe646ce3fe4ed3f0adf6919b7a50aec7daf0376" exitCode=0 Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.025895 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9lt9" event={"ID":"0f2803bc-a79f-4af4-99c7-f372b3b07bc8","Type":"ContainerDied","Data":"0ed9a2244e21941824e7de899fe646ce3fe4ed3f0adf6919b7a50aec7daf0376"} Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.051580 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.274323 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v4cn2"] Oct 14 15:18:55 crc kubenswrapper[4945]: W1014 15:18:55.283789 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63cb59cc_d19f_41e4_a839_1b67f86a5ffe.slice/crio-52c2040b26914f717cfdc3735af86e855c1bd70c385dd3e6d55bd8744a20a683 WatchSource:0}: Error finding container 52c2040b26914f717cfdc3735af86e855c1bd70c385dd3e6d55bd8744a20a683: Status 404 returned error can't find the container with id 52c2040b26914f717cfdc3735af86e855c1bd70c385dd3e6d55bd8744a20a683 Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.716670 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n87wl"] Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.718235 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.720038 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.727553 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n87wl"] Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.833982 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6gpz\" (UniqueName: \"kubernetes.io/projected/b3811f6a-94ba-4d10-bd4b-873af175eec4-kube-api-access-j6gpz\") pod \"certified-operators-n87wl\" (UID: \"b3811f6a-94ba-4d10-bd4b-873af175eec4\") " pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.834042 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3811f6a-94ba-4d10-bd4b-873af175eec4-catalog-content\") pod \"certified-operators-n87wl\" (UID: \"b3811f6a-94ba-4d10-bd4b-873af175eec4\") " pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.834141 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3811f6a-94ba-4d10-bd4b-873af175eec4-utilities\") pod \"certified-operators-n87wl\" (UID: \"b3811f6a-94ba-4d10-bd4b-873af175eec4\") " pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.935958 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3811f6a-94ba-4d10-bd4b-873af175eec4-catalog-content\") pod \"certified-operators-n87wl\" (UID: \"b3811f6a-94ba-4d10-bd4b-873af175eec4\") " pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.936282 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3811f6a-94ba-4d10-bd4b-873af175eec4-utilities\") pod \"certified-operators-n87wl\" (UID: \"b3811f6a-94ba-4d10-bd4b-873af175eec4\") " pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.936361 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6gpz\" (UniqueName: \"kubernetes.io/projected/b3811f6a-94ba-4d10-bd4b-873af175eec4-kube-api-access-j6gpz\") pod \"certified-operators-n87wl\" (UID: \"b3811f6a-94ba-4d10-bd4b-873af175eec4\") " pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.936843 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3811f6a-94ba-4d10-bd4b-873af175eec4-utilities\") pod \"certified-operators-n87wl\" (UID: \"b3811f6a-94ba-4d10-bd4b-873af175eec4\") " pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.936911 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3811f6a-94ba-4d10-bd4b-873af175eec4-catalog-content\") pod \"certified-operators-n87wl\" (UID: \"b3811f6a-94ba-4d10-bd4b-873af175eec4\") " pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:18:55 crc kubenswrapper[4945]: I1014 15:18:55.971002 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6gpz\" (UniqueName: \"kubernetes.io/projected/b3811f6a-94ba-4d10-bd4b-873af175eec4-kube-api-access-j6gpz\") pod \"certified-operators-n87wl\" (UID: \"b3811f6a-94ba-4d10-bd4b-873af175eec4\") " pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:18:56 crc kubenswrapper[4945]: I1014 15:18:56.032773 4945 generic.go:334] "Generic (PLEG): container finished" podID="63cb59cc-d19f-41e4-a839-1b67f86a5ffe" containerID="14a012e8a69427c6d65defd733074da267549e687fbfce9a23c2057473ed6031" exitCode=0 Oct 14 15:18:56 crc kubenswrapper[4945]: I1014 15:18:56.032865 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4cn2" event={"ID":"63cb59cc-d19f-41e4-a839-1b67f86a5ffe","Type":"ContainerDied","Data":"14a012e8a69427c6d65defd733074da267549e687fbfce9a23c2057473ed6031"} Oct 14 15:18:56 crc kubenswrapper[4945]: I1014 15:18:56.032949 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4cn2" event={"ID":"63cb59cc-d19f-41e4-a839-1b67f86a5ffe","Type":"ContainerStarted","Data":"52c2040b26914f717cfdc3735af86e855c1bd70c385dd3e6d55bd8744a20a683"} Oct 14 15:18:56 crc kubenswrapper[4945]: I1014 15:18:56.035443 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9lt9" event={"ID":"0f2803bc-a79f-4af4-99c7-f372b3b07bc8","Type":"ContainerStarted","Data":"72525f3267b4bb8db814d846370d5611bb92f1d708a2eaf80115c36e7d0b6213"} Oct 14 15:18:56 crc kubenswrapper[4945]: I1014 15:18:56.038674 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:18:56 crc kubenswrapper[4945]: I1014 15:18:56.069406 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m9lt9" podStartSLOduration=2.578844392 podStartE2EDuration="4.069388327s" podCreationTimestamp="2025-10-14 15:18:52 +0000 UTC" firstStartedPulling="2025-10-14 15:18:54.018829148 +0000 UTC m=+264.002877516" lastFinishedPulling="2025-10-14 15:18:55.509373063 +0000 UTC m=+265.493421451" observedRunningTime="2025-10-14 15:18:56.065262167 +0000 UTC m=+266.049310535" watchObservedRunningTime="2025-10-14 15:18:56.069388327 +0000 UTC m=+266.053436695" Oct 14 15:18:56 crc kubenswrapper[4945]: I1014 15:18:56.271771 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n87wl"] Oct 14 15:18:56 crc kubenswrapper[4945]: W1014 15:18:56.277199 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3811f6a_94ba_4d10_bd4b_873af175eec4.slice/crio-91109345b8bf767889700421d5e17426ba557827b65dd669074795fed002d656 WatchSource:0}: Error finding container 91109345b8bf767889700421d5e17426ba557827b65dd669074795fed002d656: Status 404 returned error can't find the container with id 91109345b8bf767889700421d5e17426ba557827b65dd669074795fed002d656 Oct 14 15:18:57 crc kubenswrapper[4945]: I1014 15:18:57.052262 4945 generic.go:334] "Generic (PLEG): container finished" podID="b3811f6a-94ba-4d10-bd4b-873af175eec4" containerID="a7525fbf5d89d0f12d3ce93c0c869d45f9eb0c5a4aa424c894fcedb7769460ae" exitCode=0 Oct 14 15:18:57 crc kubenswrapper[4945]: I1014 15:18:57.052304 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n87wl" event={"ID":"b3811f6a-94ba-4d10-bd4b-873af175eec4","Type":"ContainerDied","Data":"a7525fbf5d89d0f12d3ce93c0c869d45f9eb0c5a4aa424c894fcedb7769460ae"} Oct 14 15:18:57 crc kubenswrapper[4945]: I1014 15:18:57.052640 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n87wl" event={"ID":"b3811f6a-94ba-4d10-bd4b-873af175eec4","Type":"ContainerStarted","Data":"91109345b8bf767889700421d5e17426ba557827b65dd669074795fed002d656"} Oct 14 15:18:57 crc kubenswrapper[4945]: I1014 15:18:57.054481 4945 generic.go:334] "Generic (PLEG): container finished" podID="33acd0ef-71bb-467c-be47-0c93a694c8f2" containerID="72af5208e7e10ea16895a25e16d483074b37bdc90807df826c20281059c62783" exitCode=0 Oct 14 15:18:57 crc kubenswrapper[4945]: I1014 15:18:57.054538 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9z56" event={"ID":"33acd0ef-71bb-467c-be47-0c93a694c8f2","Type":"ContainerDied","Data":"72af5208e7e10ea16895a25e16d483074b37bdc90807df826c20281059c62783"} Oct 14 15:18:58 crc kubenswrapper[4945]: I1014 15:18:58.060749 4945 generic.go:334] "Generic (PLEG): container finished" podID="63cb59cc-d19f-41e4-a839-1b67f86a5ffe" containerID="d6a7be7127def6f7d20180752fecd756c2347729509c3503026a5bc8bd8b8f2a" exitCode=0 Oct 14 15:18:58 crc kubenswrapper[4945]: I1014 15:18:58.060900 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4cn2" event={"ID":"63cb59cc-d19f-41e4-a839-1b67f86a5ffe","Type":"ContainerDied","Data":"d6a7be7127def6f7d20180752fecd756c2347729509c3503026a5bc8bd8b8f2a"} Oct 14 15:18:59 crc kubenswrapper[4945]: I1014 15:18:59.068063 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9z56" event={"ID":"33acd0ef-71bb-467c-be47-0c93a694c8f2","Type":"ContainerStarted","Data":"80025c47457fde6d916a8b1249443436ec191f7756211a6e99591d1358ae9d12"} Oct 14 15:18:59 crc kubenswrapper[4945]: I1014 15:18:59.071646 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4cn2" event={"ID":"63cb59cc-d19f-41e4-a839-1b67f86a5ffe","Type":"ContainerStarted","Data":"6c5fbdee2fe5f1b48ceadd31d9e8fbc47c9cdb6c1d55a79d7a30c65ca24edd04"} Oct 14 15:18:59 crc kubenswrapper[4945]: I1014 15:18:59.073887 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n87wl" event={"ID":"b3811f6a-94ba-4d10-bd4b-873af175eec4","Type":"ContainerStarted","Data":"2a247e4a6f832c55d0e1f9d0a4bf8115fe086e8a8a1fd149a9e856d8ab53b622"} Oct 14 15:18:59 crc kubenswrapper[4945]: I1014 15:18:59.085198 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s9z56" podStartSLOduration=2.719224427 podStartE2EDuration="6.085185157s" podCreationTimestamp="2025-10-14 15:18:53 +0000 UTC" firstStartedPulling="2025-10-14 15:18:55.02571916 +0000 UTC m=+265.009767528" lastFinishedPulling="2025-10-14 15:18:58.39167985 +0000 UTC m=+268.375728258" observedRunningTime="2025-10-14 15:18:59.082395738 +0000 UTC m=+269.066444106" watchObservedRunningTime="2025-10-14 15:18:59.085185157 +0000 UTC m=+269.069233525" Oct 14 15:18:59 crc kubenswrapper[4945]: I1014 15:18:59.122804 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v4cn2" podStartSLOduration=2.403877542 podStartE2EDuration="5.122789329s" podCreationTimestamp="2025-10-14 15:18:54 +0000 UTC" firstStartedPulling="2025-10-14 15:18:56.035843854 +0000 UTC m=+266.019892222" lastFinishedPulling="2025-10-14 15:18:58.754755641 +0000 UTC m=+268.738804009" observedRunningTime="2025-10-14 15:18:59.121161867 +0000 UTC m=+269.105210245" watchObservedRunningTime="2025-10-14 15:18:59.122789329 +0000 UTC m=+269.106837697" Oct 14 15:19:00 crc kubenswrapper[4945]: I1014 15:19:00.081571 4945 generic.go:334] "Generic (PLEG): container finished" podID="b3811f6a-94ba-4d10-bd4b-873af175eec4" containerID="2a247e4a6f832c55d0e1f9d0a4bf8115fe086e8a8a1fd149a9e856d8ab53b622" exitCode=0 Oct 14 15:19:00 crc kubenswrapper[4945]: I1014 15:19:00.082484 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n87wl" event={"ID":"b3811f6a-94ba-4d10-bd4b-873af175eec4","Type":"ContainerDied","Data":"2a247e4a6f832c55d0e1f9d0a4bf8115fe086e8a8a1fd149a9e856d8ab53b622"} Oct 14 15:19:01 crc kubenswrapper[4945]: I1014 15:19:01.088651 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n87wl" event={"ID":"b3811f6a-94ba-4d10-bd4b-873af175eec4","Type":"ContainerStarted","Data":"ed8b248b128314b208cd5836cc79025ce58f554e100fc87e8bfa6dbb75d22042"} Oct 14 15:19:01 crc kubenswrapper[4945]: I1014 15:19:01.107163 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n87wl" podStartSLOduration=2.686231706 podStartE2EDuration="6.107147109s" podCreationTimestamp="2025-10-14 15:18:55 +0000 UTC" firstStartedPulling="2025-10-14 15:18:57.111482584 +0000 UTC m=+267.095530992" lastFinishedPulling="2025-10-14 15:19:00.532397987 +0000 UTC m=+270.516446395" observedRunningTime="2025-10-14 15:19:01.104860996 +0000 UTC m=+271.088909364" watchObservedRunningTime="2025-10-14 15:19:01.107147109 +0000 UTC m=+271.091195477" Oct 14 15:19:02 crc kubenswrapper[4945]: I1014 15:19:02.637559 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:19:02 crc kubenswrapper[4945]: I1014 15:19:02.637923 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:19:02 crc kubenswrapper[4945]: I1014 15:19:02.673021 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:19:03 crc kubenswrapper[4945]: I1014 15:19:03.155483 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m9lt9" Oct 14 15:19:03 crc kubenswrapper[4945]: I1014 15:19:03.651531 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:19:03 crc kubenswrapper[4945]: I1014 15:19:03.651904 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:19:03 crc kubenswrapper[4945]: I1014 15:19:03.717587 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:19:04 crc kubenswrapper[4945]: I1014 15:19:04.144709 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s9z56" Oct 14 15:19:04 crc kubenswrapper[4945]: I1014 15:19:04.884228 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-mq2vp" Oct 14 15:19:04 crc kubenswrapper[4945]: I1014 15:19:04.928299 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4bf5h"] Oct 14 15:19:05 crc kubenswrapper[4945]: I1014 15:19:05.052381 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:19:05 crc kubenswrapper[4945]: I1014 15:19:05.052622 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:19:05 crc kubenswrapper[4945]: I1014 15:19:05.087763 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:19:05 crc kubenswrapper[4945]: I1014 15:19:05.150958 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v4cn2" Oct 14 15:19:06 crc kubenswrapper[4945]: I1014 15:19:06.039845 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:19:06 crc kubenswrapper[4945]: I1014 15:19:06.040901 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:19:06 crc kubenswrapper[4945]: I1014 15:19:06.084391 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:19:06 crc kubenswrapper[4945]: I1014 15:19:06.153615 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:19:15 crc kubenswrapper[4945]: I1014 15:19:15.503741 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" podUID="d094c180-e6dd-4064-aea1-4c04b43e4963" containerName="oauth-openshift" containerID="cri-o://c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337" gracePeriod=15 Oct 14 15:19:15 crc kubenswrapper[4945]: I1014 15:19:15.956574 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:19:15 crc kubenswrapper[4945]: I1014 15:19:15.988178 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-59cf6c497f-d5vlm"] Oct 14 15:19:15 crc kubenswrapper[4945]: E1014 15:19:15.988437 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d094c180-e6dd-4064-aea1-4c04b43e4963" containerName="oauth-openshift" Oct 14 15:19:15 crc kubenswrapper[4945]: I1014 15:19:15.988455 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d094c180-e6dd-4064-aea1-4c04b43e4963" containerName="oauth-openshift" Oct 14 15:19:15 crc kubenswrapper[4945]: I1014 15:19:15.988570 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d094c180-e6dd-4064-aea1-4c04b43e4963" containerName="oauth-openshift" Oct 14 15:19:15 crc kubenswrapper[4945]: I1014 15:19:15.990147 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.004177 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-59cf6c497f-d5vlm"] Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.026925 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-trusted-ca-bundle\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027012 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7661ece-ee26-4c7e-82fc-884abe37b32d-audit-policies\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027034 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltzg8\" (UniqueName: \"kubernetes.io/projected/b7661ece-ee26-4c7e-82fc-884abe37b32d-kube-api-access-ltzg8\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027053 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7661ece-ee26-4c7e-82fc-884abe37b32d-audit-dir\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027069 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-user-template-error\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027095 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027123 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-service-ca\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027152 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027177 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027196 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027229 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027251 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-router-certs\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027279 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-session\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027320 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.027344 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-user-template-login\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.028238 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128238 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-provider-selection\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128601 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-audit-policies\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128633 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-router-certs\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128660 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-ocp-branding-template\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128709 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-service-ca\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128734 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-error\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128772 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-session\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128795 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pztn\" (UniqueName: \"kubernetes.io/projected/d094c180-e6dd-4064-aea1-4c04b43e4963-kube-api-access-2pztn\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128823 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-login\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128853 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-cliconfig\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128905 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-idp-0-file-data\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128927 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d094c180-e6dd-4064-aea1-4c04b43e4963-audit-dir\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.128948 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-serving-cert\") pod \"d094c180-e6dd-4064-aea1-4c04b43e4963\" (UID: \"d094c180-e6dd-4064-aea1-4c04b43e4963\") " Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129075 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129108 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-router-certs\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129153 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-session\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129178 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-user-template-login\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129204 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129227 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7661ece-ee26-4c7e-82fc-884abe37b32d-audit-policies\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129250 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltzg8\" (UniqueName: \"kubernetes.io/projected/b7661ece-ee26-4c7e-82fc-884abe37b32d-kube-api-access-ltzg8\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129277 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7661ece-ee26-4c7e-82fc-884abe37b32d-audit-dir\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129300 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-user-template-error\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129331 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129355 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-service-ca\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129391 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129424 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129454 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129500 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129570 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.129851 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.131225 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b7661ece-ee26-4c7e-82fc-884abe37b32d-audit-dir\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.131695 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b7661ece-ee26-4c7e-82fc-884abe37b32d-audit-policies\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.132132 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.132270 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-service-ca\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.132549 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.133187 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.133405 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d094c180-e6dd-4064-aea1-4c04b43e4963-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.135334 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.135578 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.136511 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.137119 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.137234 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.137250 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.137373 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-router-certs\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.137519 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-user-template-login\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.138178 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-user-template-error\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.139025 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.139207 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.139501 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.139738 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.140231 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.140365 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.143274 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b7661ece-ee26-4c7e-82fc-884abe37b32d-v4-0-config-system-session\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.145511 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d094c180-e6dd-4064-aea1-4c04b43e4963-kube-api-access-2pztn" (OuterVolumeSpecName: "kube-api-access-2pztn") pod "d094c180-e6dd-4064-aea1-4c04b43e4963" (UID: "d094c180-e6dd-4064-aea1-4c04b43e4963"). InnerVolumeSpecName "kube-api-access-2pztn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.150834 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltzg8\" (UniqueName: \"kubernetes.io/projected/b7661ece-ee26-4c7e-82fc-884abe37b32d-kube-api-access-ltzg8\") pod \"oauth-openshift-59cf6c497f-d5vlm\" (UID: \"b7661ece-ee26-4c7e-82fc-884abe37b32d\") " pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.177334 4945 generic.go:334] "Generic (PLEG): container finished" podID="d094c180-e6dd-4064-aea1-4c04b43e4963" containerID="c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337" exitCode=0 Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.177370 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.177395 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" event={"ID":"d094c180-e6dd-4064-aea1-4c04b43e4963","Type":"ContainerDied","Data":"c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337"} Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.178226 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9zzcq" event={"ID":"d094c180-e6dd-4064-aea1-4c04b43e4963","Type":"ContainerDied","Data":"3b88c808254ceeb048e67aa5549253f449702b2188e3500c514336d6fbc575b2"} Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.178247 4945 scope.go:117] "RemoveContainer" containerID="c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.211124 4945 scope.go:117] "RemoveContainer" containerID="c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337" Oct 14 15:19:16 crc kubenswrapper[4945]: E1014 15:19:16.211584 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337\": container with ID starting with c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337 not found: ID does not exist" containerID="c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.211632 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337"} err="failed to get container status \"c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337\": rpc error: code = NotFound desc = could not find container \"c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337\": container with ID starting with c5ee1cfce381bbb391565fce0d59171362ac654cac7aee98df25f5d82472e337 not found: ID does not exist" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.219773 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9zzcq"] Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.223772 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9zzcq"] Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230417 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230467 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230481 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230495 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pztn\" (UniqueName: \"kubernetes.io/projected/d094c180-e6dd-4064-aea1-4c04b43e4963-kube-api-access-2pztn\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230506 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230521 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230532 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230543 4945 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d094c180-e6dd-4064-aea1-4c04b43e4963-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230555 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230568 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230583 4945 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d094c180-e6dd-4064-aea1-4c04b43e4963-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230594 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.230606 4945 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d094c180-e6dd-4064-aea1-4c04b43e4963-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.306325 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.741182 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-59cf6c497f-d5vlm"] Oct 14 15:19:16 crc kubenswrapper[4945]: W1014 15:19:16.745285 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7661ece_ee26_4c7e_82fc_884abe37b32d.slice/crio-5fd1b4d191f9bb09c936a3bdb1c7c6f222dd170e3e2c28672f60e978f3774fc2 WatchSource:0}: Error finding container 5fd1b4d191f9bb09c936a3bdb1c7c6f222dd170e3e2c28672f60e978f3774fc2: Status 404 returned error can't find the container with id 5fd1b4d191f9bb09c936a3bdb1c7c6f222dd170e3e2c28672f60e978f3774fc2 Oct 14 15:19:16 crc kubenswrapper[4945]: I1014 15:19:16.771060 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d094c180-e6dd-4064-aea1-4c04b43e4963" path="/var/lib/kubelet/pods/d094c180-e6dd-4064-aea1-4c04b43e4963/volumes" Oct 14 15:19:17 crc kubenswrapper[4945]: I1014 15:19:17.186388 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" event={"ID":"b7661ece-ee26-4c7e-82fc-884abe37b32d","Type":"ContainerStarted","Data":"aaec3fb0828815425923ba70021ddf867626556ee20b7aa134289b142de7779f"} Oct 14 15:19:17 crc kubenswrapper[4945]: I1014 15:19:17.186447 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" event={"ID":"b7661ece-ee26-4c7e-82fc-884abe37b32d","Type":"ContainerStarted","Data":"5fd1b4d191f9bb09c936a3bdb1c7c6f222dd170e3e2c28672f60e978f3774fc2"} Oct 14 15:19:17 crc kubenswrapper[4945]: I1014 15:19:17.186730 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:17 crc kubenswrapper[4945]: I1014 15:19:17.233850 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" podStartSLOduration=27.233814137 podStartE2EDuration="27.233814137s" podCreationTimestamp="2025-10-14 15:18:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:19:17.223997801 +0000 UTC m=+287.208046209" watchObservedRunningTime="2025-10-14 15:19:17.233814137 +0000 UTC m=+287.217862545" Oct 14 15:19:17 crc kubenswrapper[4945]: I1014 15:19:17.318002 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-59cf6c497f-d5vlm" Oct 14 15:19:29 crc kubenswrapper[4945]: I1014 15:19:29.968637 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" podUID="09ab4e69-a6c2-41f6-a1e6-d59919746fd9" containerName="registry" containerID="cri-o://24e5d099d02a577a3bee77f291fa8fc870b73ecc8d22747d76e85c70d8a8482c" gracePeriod=30 Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.266439 4945 generic.go:334] "Generic (PLEG): container finished" podID="09ab4e69-a6c2-41f6-a1e6-d59919746fd9" containerID="24e5d099d02a577a3bee77f291fa8fc870b73ecc8d22747d76e85c70d8a8482c" exitCode=0 Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.266512 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" event={"ID":"09ab4e69-a6c2-41f6-a1e6-d59919746fd9","Type":"ContainerDied","Data":"24e5d099d02a577a3bee77f291fa8fc870b73ecc8d22747d76e85c70d8a8482c"} Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.337212 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.439657 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-registry-certificates\") pod \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.439712 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-registry-tls\") pod \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.439740 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-trusted-ca\") pod \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.439760 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c26cc\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-kube-api-access-c26cc\") pod \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.439822 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-bound-sa-token\") pod \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.439847 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-installation-pull-secrets\") pod \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.440190 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.440246 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-ca-trust-extracted\") pod \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\" (UID: \"09ab4e69-a6c2-41f6-a1e6-d59919746fd9\") " Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.441058 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "09ab4e69-a6c2-41f6-a1e6-d59919746fd9" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.441157 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "09ab4e69-a6c2-41f6-a1e6-d59919746fd9" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.449855 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-kube-api-access-c26cc" (OuterVolumeSpecName: "kube-api-access-c26cc") pod "09ab4e69-a6c2-41f6-a1e6-d59919746fd9" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9"). InnerVolumeSpecName "kube-api-access-c26cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.450448 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "09ab4e69-a6c2-41f6-a1e6-d59919746fd9" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.451016 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "09ab4e69-a6c2-41f6-a1e6-d59919746fd9" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.455178 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "09ab4e69-a6c2-41f6-a1e6-d59919746fd9" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.456091 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "09ab4e69-a6c2-41f6-a1e6-d59919746fd9" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.463549 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "09ab4e69-a6c2-41f6-a1e6-d59919746fd9" (UID: "09ab4e69-a6c2-41f6-a1e6-d59919746fd9"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.541269 4945 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.541306 4945 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.541323 4945 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.541333 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.541346 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c26cc\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-kube-api-access-c26cc\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.541359 4945 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:30 crc kubenswrapper[4945]: I1014 15:19:30.541374 4945 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/09ab4e69-a6c2-41f6-a1e6-d59919746fd9-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 15:19:31 crc kubenswrapper[4945]: I1014 15:19:31.284513 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" event={"ID":"09ab4e69-a6c2-41f6-a1e6-d59919746fd9","Type":"ContainerDied","Data":"48d54052f7678f8046e4f38ae0044d168dc1891b5d867e7c2e8d4ebc001bcfc2"} Oct 14 15:19:31 crc kubenswrapper[4945]: I1014 15:19:31.284620 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4bf5h" Oct 14 15:19:31 crc kubenswrapper[4945]: I1014 15:19:31.284801 4945 scope.go:117] "RemoveContainer" containerID="24e5d099d02a577a3bee77f291fa8fc870b73ecc8d22747d76e85c70d8a8482c" Oct 14 15:19:31 crc kubenswrapper[4945]: I1014 15:19:31.331039 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4bf5h"] Oct 14 15:19:31 crc kubenswrapper[4945]: I1014 15:19:31.335826 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4bf5h"] Oct 14 15:19:32 crc kubenswrapper[4945]: I1014 15:19:32.784289 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ab4e69-a6c2-41f6-a1e6-d59919746fd9" path="/var/lib/kubelet/pods/09ab4e69-a6c2-41f6-a1e6-d59919746fd9/volumes" Oct 14 15:20:46 crc kubenswrapper[4945]: I1014 15:20:46.795601 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:20:46 crc kubenswrapper[4945]: I1014 15:20:46.797105 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:21:16 crc kubenswrapper[4945]: I1014 15:21:16.796253 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:21:16 crc kubenswrapper[4945]: I1014 15:21:16.796806 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:21:46 crc kubenswrapper[4945]: I1014 15:21:46.795411 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:21:46 crc kubenswrapper[4945]: I1014 15:21:46.796131 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:21:46 crc kubenswrapper[4945]: I1014 15:21:46.796176 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:21:46 crc kubenswrapper[4945]: I1014 15:21:46.796800 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c8c34e87ab2d9cc333cc18624592bb9519696247cdd2c0458ada848829da5a55"} pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 15:21:46 crc kubenswrapper[4945]: I1014 15:21:46.796866 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" containerID="cri-o://c8c34e87ab2d9cc333cc18624592bb9519696247cdd2c0458ada848829da5a55" gracePeriod=600 Oct 14 15:21:47 crc kubenswrapper[4945]: I1014 15:21:47.105092 4945 generic.go:334] "Generic (PLEG): container finished" podID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerID="c8c34e87ab2d9cc333cc18624592bb9519696247cdd2c0458ada848829da5a55" exitCode=0 Oct 14 15:21:47 crc kubenswrapper[4945]: I1014 15:21:47.105144 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerDied","Data":"c8c34e87ab2d9cc333cc18624592bb9519696247cdd2c0458ada848829da5a55"} Oct 14 15:21:47 crc kubenswrapper[4945]: I1014 15:21:47.105462 4945 scope.go:117] "RemoveContainer" containerID="e68d0ab4b1270aa4108915c476f689708c9d309c53fc6fb753b955ebbac963dd" Oct 14 15:21:48 crc kubenswrapper[4945]: I1014 15:21:48.111394 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerStarted","Data":"8f32c2d032f723cbbb763494c9ff11fd408a6db0a40ec4819ef209acfe780ed8"} Oct 14 15:23:30 crc kubenswrapper[4945]: I1014 15:23:30.918814 4945 scope.go:117] "RemoveContainer" containerID="a6150ee7d76eb524532781a91fe5de51725e6b65e0319f922a369bb556d79c74" Oct 14 15:24:16 crc kubenswrapper[4945]: I1014 15:24:16.795725 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:24:16 crc kubenswrapper[4945]: I1014 15:24:16.796501 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:24:30 crc kubenswrapper[4945]: I1014 15:24:30.959918 4945 scope.go:117] "RemoveContainer" containerID="f4e648faedb9ce11a928967ed75d54653e1567161c49af283f14f824ecc991d4" Oct 14 15:24:30 crc kubenswrapper[4945]: I1014 15:24:30.991028 4945 scope.go:117] "RemoveContainer" containerID="97e4d44348b3e6dea11ce84bee2e6afadb6b6534f3e24235f30d7c34fc04f864" Oct 14 15:24:46 crc kubenswrapper[4945]: I1014 15:24:46.795614 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:24:46 crc kubenswrapper[4945]: I1014 15:24:46.796193 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.842722 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-kw8kj"] Oct 14 15:25:04 crc kubenswrapper[4945]: E1014 15:25:04.843520 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ab4e69-a6c2-41f6-a1e6-d59919746fd9" containerName="registry" Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.843533 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ab4e69-a6c2-41f6-a1e6-d59919746fd9" containerName="registry" Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.843651 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="09ab4e69-a6c2-41f6-a1e6-d59919746fd9" containerName="registry" Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.844188 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-kw8kj" Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.845859 4945 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-clwcd" Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.846078 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.846151 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.849218 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-kw8kj"] Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.888149 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-jmvq2"] Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.889030 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-jmvq2" Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.892351 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-j5ssh"] Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.893227 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-j5ssh" Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.895249 4945 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-55ttq" Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.895550 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-jmvq2"] Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.900933 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-j5ssh"] Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.909489 4945 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-r5889" Oct 14 15:25:04 crc kubenswrapper[4945]: I1014 15:25:04.983689 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nll9f\" (UniqueName: \"kubernetes.io/projected/2385f0e3-9e8d-4971-95e6-08dbf4af170c-kube-api-access-nll9f\") pod \"cert-manager-cainjector-7f985d654d-kw8kj\" (UID: \"2385f0e3-9e8d-4971-95e6-08dbf4af170c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-kw8kj" Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.085032 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nll9f\" (UniqueName: \"kubernetes.io/projected/2385f0e3-9e8d-4971-95e6-08dbf4af170c-kube-api-access-nll9f\") pod \"cert-manager-cainjector-7f985d654d-kw8kj\" (UID: \"2385f0e3-9e8d-4971-95e6-08dbf4af170c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-kw8kj" Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.085114 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2rsn\" (UniqueName: \"kubernetes.io/projected/6a2a312b-2002-4494-b432-efb6060b5a89-kube-api-access-v2rsn\") pod \"cert-manager-5b446d88c5-jmvq2\" (UID: \"6a2a312b-2002-4494-b432-efb6060b5a89\") " pod="cert-manager/cert-manager-5b446d88c5-jmvq2" Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.085167 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8hfr\" (UniqueName: \"kubernetes.io/projected/a59d5293-e44c-459d-9875-137a289aa9d8-kube-api-access-g8hfr\") pod \"cert-manager-webhook-5655c58dd6-j5ssh\" (UID: \"a59d5293-e44c-459d-9875-137a289aa9d8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-j5ssh" Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.109803 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nll9f\" (UniqueName: \"kubernetes.io/projected/2385f0e3-9e8d-4971-95e6-08dbf4af170c-kube-api-access-nll9f\") pod \"cert-manager-cainjector-7f985d654d-kw8kj\" (UID: \"2385f0e3-9e8d-4971-95e6-08dbf4af170c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-kw8kj" Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.170157 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-kw8kj" Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.186929 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2rsn\" (UniqueName: \"kubernetes.io/projected/6a2a312b-2002-4494-b432-efb6060b5a89-kube-api-access-v2rsn\") pod \"cert-manager-5b446d88c5-jmvq2\" (UID: \"6a2a312b-2002-4494-b432-efb6060b5a89\") " pod="cert-manager/cert-manager-5b446d88c5-jmvq2" Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.186988 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8hfr\" (UniqueName: \"kubernetes.io/projected/a59d5293-e44c-459d-9875-137a289aa9d8-kube-api-access-g8hfr\") pod \"cert-manager-webhook-5655c58dd6-j5ssh\" (UID: \"a59d5293-e44c-459d-9875-137a289aa9d8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-j5ssh" Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.218649 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2rsn\" (UniqueName: \"kubernetes.io/projected/6a2a312b-2002-4494-b432-efb6060b5a89-kube-api-access-v2rsn\") pod \"cert-manager-5b446d88c5-jmvq2\" (UID: \"6a2a312b-2002-4494-b432-efb6060b5a89\") " pod="cert-manager/cert-manager-5b446d88c5-jmvq2" Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.218829 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8hfr\" (UniqueName: \"kubernetes.io/projected/a59d5293-e44c-459d-9875-137a289aa9d8-kube-api-access-g8hfr\") pod \"cert-manager-webhook-5655c58dd6-j5ssh\" (UID: \"a59d5293-e44c-459d-9875-137a289aa9d8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-j5ssh" Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.366462 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-kw8kj"] Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.373232 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.402057 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-kw8kj" event={"ID":"2385f0e3-9e8d-4971-95e6-08dbf4af170c","Type":"ContainerStarted","Data":"bb0ec52470ebab74394b519557400a3b56a4831ba7fa3e52381a809bc64624c5"} Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.508746 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-jmvq2" Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.516405 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-j5ssh" Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.712252 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-j5ssh"] Oct 14 15:25:05 crc kubenswrapper[4945]: W1014 15:25:05.717147 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda59d5293_e44c_459d_9875_137a289aa9d8.slice/crio-e3543a0b1f123b51a80a199007379803883eed18e5e49e3eb4bc766b03321289 WatchSource:0}: Error finding container e3543a0b1f123b51a80a199007379803883eed18e5e49e3eb4bc766b03321289: Status 404 returned error can't find the container with id e3543a0b1f123b51a80a199007379803883eed18e5e49e3eb4bc766b03321289 Oct 14 15:25:05 crc kubenswrapper[4945]: I1014 15:25:05.749430 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-jmvq2"] Oct 14 15:25:06 crc kubenswrapper[4945]: I1014 15:25:06.409135 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-jmvq2" event={"ID":"6a2a312b-2002-4494-b432-efb6060b5a89","Type":"ContainerStarted","Data":"bc699451ecedac9c25884925376899bf6dd7cbe8db0753350bf278a032d52b35"} Oct 14 15:25:06 crc kubenswrapper[4945]: I1014 15:25:06.410337 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-j5ssh" event={"ID":"a59d5293-e44c-459d-9875-137a289aa9d8","Type":"ContainerStarted","Data":"e3543a0b1f123b51a80a199007379803883eed18e5e49e3eb4bc766b03321289"} Oct 14 15:25:09 crc kubenswrapper[4945]: I1014 15:25:09.426034 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-j5ssh" event={"ID":"a59d5293-e44c-459d-9875-137a289aa9d8","Type":"ContainerStarted","Data":"27023605d445b9ffe4bbe7d5199ff742c7c11bc9f3ee68b32352f173b4e9ed43"} Oct 14 15:25:09 crc kubenswrapper[4945]: I1014 15:25:09.426662 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-j5ssh" Oct 14 15:25:09 crc kubenswrapper[4945]: I1014 15:25:09.427167 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-kw8kj" event={"ID":"2385f0e3-9e8d-4971-95e6-08dbf4af170c","Type":"ContainerStarted","Data":"80e063a9cad9d9a8ef4f7097c6b00b99dd59a9a40daf084ad44ea1e5072eb7dd"} Oct 14 15:25:09 crc kubenswrapper[4945]: I1014 15:25:09.444705 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-j5ssh" podStartSLOduration=2.082911324 podStartE2EDuration="5.444686569s" podCreationTimestamp="2025-10-14 15:25:04 +0000 UTC" firstStartedPulling="2025-10-14 15:25:05.719456565 +0000 UTC m=+635.703504933" lastFinishedPulling="2025-10-14 15:25:09.08123177 +0000 UTC m=+639.065280178" observedRunningTime="2025-10-14 15:25:09.443443354 +0000 UTC m=+639.427491752" watchObservedRunningTime="2025-10-14 15:25:09.444686569 +0000 UTC m=+639.428734947" Oct 14 15:25:10 crc kubenswrapper[4945]: I1014 15:25:10.434504 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-jmvq2" event={"ID":"6a2a312b-2002-4494-b432-efb6060b5a89","Type":"ContainerStarted","Data":"6e91c7e51fe2273dbd069d6a87de32f09c76ff933547674f05cecd50f84e4877"} Oct 14 15:25:10 crc kubenswrapper[4945]: I1014 15:25:10.455564 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-jmvq2" podStartSLOduration=2.8543885959999997 podStartE2EDuration="6.455545183s" podCreationTimestamp="2025-10-14 15:25:04 +0000 UTC" firstStartedPulling="2025-10-14 15:25:05.761308598 +0000 UTC m=+635.745356966" lastFinishedPulling="2025-10-14 15:25:09.362465185 +0000 UTC m=+639.346513553" observedRunningTime="2025-10-14 15:25:10.453301139 +0000 UTC m=+640.437349517" watchObservedRunningTime="2025-10-14 15:25:10.455545183 +0000 UTC m=+640.439593551" Oct 14 15:25:10 crc kubenswrapper[4945]: I1014 15:25:10.456357 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-kw8kj" podStartSLOduration=2.739587865 podStartE2EDuration="6.456350646s" podCreationTimestamp="2025-10-14 15:25:04 +0000 UTC" firstStartedPulling="2025-10-14 15:25:05.373043475 +0000 UTC m=+635.357091843" lastFinishedPulling="2025-10-14 15:25:09.089806256 +0000 UTC m=+639.073854624" observedRunningTime="2025-10-14 15:25:09.481768545 +0000 UTC m=+639.465816913" watchObservedRunningTime="2025-10-14 15:25:10.456350646 +0000 UTC m=+640.440399014" Oct 14 15:25:14 crc kubenswrapper[4945]: I1014 15:25:14.877492 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7pqtt"] Oct 14 15:25:14 crc kubenswrapper[4945]: I1014 15:25:14.877970 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovn-controller" containerID="cri-o://e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488" gracePeriod=30 Oct 14 15:25:14 crc kubenswrapper[4945]: I1014 15:25:14.878025 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="nbdb" containerID="cri-o://251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b" gracePeriod=30 Oct 14 15:25:14 crc kubenswrapper[4945]: I1014 15:25:14.878111 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="northd" containerID="cri-o://3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d" gracePeriod=30 Oct 14 15:25:14 crc kubenswrapper[4945]: I1014 15:25:14.878160 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59" gracePeriod=30 Oct 14 15:25:14 crc kubenswrapper[4945]: I1014 15:25:14.878203 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="kube-rbac-proxy-node" containerID="cri-o://85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563" gracePeriod=30 Oct 14 15:25:14 crc kubenswrapper[4945]: I1014 15:25:14.878243 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovn-acl-logging" containerID="cri-o://7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04" gracePeriod=30 Oct 14 15:25:14 crc kubenswrapper[4945]: I1014 15:25:14.878306 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="sbdb" containerID="cri-o://228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0" gracePeriod=30 Oct 14 15:25:14 crc kubenswrapper[4945]: I1014 15:25:14.919472 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" containerID="cri-o://b7e9738cf3f80c415296bbc75a8735490f3d6112c9285ca52958a2fc67bda65e" gracePeriod=30 Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.473534 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b76xs_f82b9cd9-7a14-4dd9-bb0e-fde942389666/kube-multus/2.log" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.474279 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b76xs_f82b9cd9-7a14-4dd9-bb0e-fde942389666/kube-multus/1.log" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.474409 4945 generic.go:334] "Generic (PLEG): container finished" podID="f82b9cd9-7a14-4dd9-bb0e-fde942389666" containerID="6daedef21246bc41a36b48797081a15d5a44d4a71e4e65c78f31574913dcdb57" exitCode=2 Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.474526 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b76xs" event={"ID":"f82b9cd9-7a14-4dd9-bb0e-fde942389666","Type":"ContainerDied","Data":"6daedef21246bc41a36b48797081a15d5a44d4a71e4e65c78f31574913dcdb57"} Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.474613 4945 scope.go:117] "RemoveContainer" containerID="17eb13e41970758fd540b4fff10d1848e4f0da616768856063641a4d1994340c" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.475119 4945 scope.go:117] "RemoveContainer" containerID="6daedef21246bc41a36b48797081a15d5a44d4a71e4e65c78f31574913dcdb57" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.475415 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-b76xs_openshift-multus(f82b9cd9-7a14-4dd9-bb0e-fde942389666)\"" pod="openshift-multus/multus-b76xs" podUID="f82b9cd9-7a14-4dd9-bb0e-fde942389666" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.482652 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovnkube-controller/3.log" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.485061 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovn-acl-logging/0.log" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.485560 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovn-controller/0.log" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486305 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="b7e9738cf3f80c415296bbc75a8735490f3d6112c9285ca52958a2fc67bda65e" exitCode=0 Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486344 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0" exitCode=0 Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486356 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b" exitCode=0 Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486368 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d" exitCode=0 Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486379 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59" exitCode=0 Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486390 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563" exitCode=0 Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486401 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04" exitCode=143 Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486413 4945 generic.go:334] "Generic (PLEG): container finished" podID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerID="e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488" exitCode=143 Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486439 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"b7e9738cf3f80c415296bbc75a8735490f3d6112c9285ca52958a2fc67bda65e"} Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486471 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0"} Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486488 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b"} Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486503 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d"} Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486519 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59"} Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486533 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563"} Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486545 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04"} Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.486557 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488"} Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.520321 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-j5ssh" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.537637 4945 scope.go:117] "RemoveContainer" containerID="70cbee7f317255e6bd15d353fa4cb05fd05cd5524922a8c2aa717cb1be2b3134" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.686359 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovn-acl-logging/0.log" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.688044 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovn-controller/0.log" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.688900 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.742792 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6l8mm"] Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743276 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743319 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743327 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743333 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743340 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743346 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743361 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovn-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743366 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovn-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743399 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="kube-rbac-proxy-node" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743409 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="kube-rbac-proxy-node" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743421 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="kube-rbac-proxy-ovn-metrics" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743429 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="kube-rbac-proxy-ovn-metrics" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743438 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="sbdb" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743445 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="sbdb" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743480 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="kubecfg-setup" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743490 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="kubecfg-setup" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743500 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="northd" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743507 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="northd" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743517 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovn-acl-logging" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743523 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovn-acl-logging" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743529 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="nbdb" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743555 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="nbdb" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743669 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743678 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="kube-rbac-proxy-node" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743687 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="sbdb" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743714 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743722 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743730 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovn-acl-logging" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743736 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovn-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743744 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="nbdb" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743752 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="northd" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743758 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743766 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="kube-rbac-proxy-ovn-metrics" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743907 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743915 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: E1014 15:25:15.743923 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.743928 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.744057 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" containerName="ovnkube-controller" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.746012 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833025 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovn-node-metrics-cert\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833163 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-run-netns\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833238 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833277 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-ovn\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833353 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833458 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-cni-bin\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833474 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-cni-netd\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833511 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833521 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-etc-openvswitch\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833536 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833547 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovnkube-script-lib\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833555 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833578 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-var-lib-openvswitch\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833594 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-kubelet\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833614 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833638 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-openvswitch\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833642 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833675 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62tc9\" (UniqueName: \"kubernetes.io/projected/6de8ddf6-38db-4394-8660-7e9659bf2bd2-kube-api-access-62tc9\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833705 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-slash\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833725 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-systemd\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833745 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-systemd-units\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833764 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-env-overrides\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833783 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-node-log\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833798 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-log-socket\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833827 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovnkube-config\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.833858 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-run-ovn-kubernetes\") pod \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\" (UID: \"6de8ddf6-38db-4394-8660-7e9659bf2bd2\") " Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834001 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-var-lib-openvswitch\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834033 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-run-netns\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834052 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-ovnkube-config\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834077 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-node-log\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834108 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-env-overrides\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834139 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834158 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp6qf\" (UniqueName: \"kubernetes.io/projected/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-kube-api-access-fp6qf\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834178 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-run-openvswitch\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834204 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-ovnkube-script-lib\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834236 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-kubelet\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834266 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-cni-bin\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834287 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-etc-openvswitch\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834317 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-run-systemd\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834334 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-cni-netd\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834353 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-run-ovn-kubernetes\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834372 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-slash\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834391 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-run-ovn\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834407 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-log-socket\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834426 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-ovn-node-metrics-cert\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834449 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-systemd-units\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834487 4945 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834497 4945 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834506 4945 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834514 4945 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834522 4945 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834530 4945 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834032 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834051 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834064 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834079 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834092 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834710 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834766 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834805 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-node-log" (OuterVolumeSpecName: "node-log") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834829 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-log-socket" (OuterVolumeSpecName: "log-socket") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.834854 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-slash" (OuterVolumeSpecName: "host-slash") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.835049 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.840022 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6de8ddf6-38db-4394-8660-7e9659bf2bd2-kube-api-access-62tc9" (OuterVolumeSpecName: "kube-api-access-62tc9") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "kube-api-access-62tc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.840406 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.853620 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "6de8ddf6-38db-4394-8660-7e9659bf2bd2" (UID: "6de8ddf6-38db-4394-8660-7e9659bf2bd2"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.935340 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-run-netns\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.935411 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-ovnkube-config\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.935450 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-node-log\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.935492 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-env-overrides\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.935551 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.935579 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-run-netns\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.935664 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-node-log\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.935590 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp6qf\" (UniqueName: \"kubernetes.io/projected/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-kube-api-access-fp6qf\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936027 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-run-openvswitch\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936127 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-run-openvswitch\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936248 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-ovnkube-script-lib\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936308 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-ovnkube-config\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936387 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-kubelet\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936437 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-cni-bin\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936479 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-kubelet\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936500 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-etc-openvswitch\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936524 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-env-overrides\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936509 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-cni-bin\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936614 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-etc-openvswitch\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936654 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936707 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-run-systemd\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936774 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-cni-netd\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936779 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-run-systemd\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936836 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-run-ovn-kubernetes\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936834 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-cni-netd\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936871 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-run-ovn-kubernetes\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.936922 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-slash\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937033 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-host-slash\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937006 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-run-ovn\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937071 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-ovnkube-script-lib\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937081 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-run-ovn\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937132 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-log-socket\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937198 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-log-socket\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937211 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-ovn-node-metrics-cert\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937314 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-systemd-units\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937424 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-systemd-units\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937525 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-var-lib-openvswitch\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937632 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-var-lib-openvswitch\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937828 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937866 4945 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937932 4945 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937960 4945 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.937985 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62tc9\" (UniqueName: \"kubernetes.io/projected/6de8ddf6-38db-4394-8660-7e9659bf2bd2-kube-api-access-62tc9\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.938009 4945 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-slash\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.938065 4945 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.938083 4945 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.938099 4945 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.938116 4945 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-node-log\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.938132 4945 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-log-socket\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.938148 4945 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.938165 4945 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6de8ddf6-38db-4394-8660-7e9659bf2bd2-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.938182 4945 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6de8ddf6-38db-4394-8660-7e9659bf2bd2-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.941659 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-ovn-node-metrics-cert\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:15 crc kubenswrapper[4945]: I1014 15:25:15.953352 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp6qf\" (UniqueName: \"kubernetes.io/projected/258a4e61-6fdc-495b-9f6d-fffb6e3ee702-kube-api-access-fp6qf\") pod \"ovnkube-node-6l8mm\" (UID: \"258a4e61-6fdc-495b-9f6d-fffb6e3ee702\") " pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.066005 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.494119 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b76xs_f82b9cd9-7a14-4dd9-bb0e-fde942389666/kube-multus/2.log" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.496572 4945 generic.go:334] "Generic (PLEG): container finished" podID="258a4e61-6fdc-495b-9f6d-fffb6e3ee702" containerID="1e2265c1ddc0101818e6ba023ad8cc0c82cecd207621f82f40cdad9106f0d6f0" exitCode=0 Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.496649 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" event={"ID":"258a4e61-6fdc-495b-9f6d-fffb6e3ee702","Type":"ContainerDied","Data":"1e2265c1ddc0101818e6ba023ad8cc0c82cecd207621f82f40cdad9106f0d6f0"} Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.496677 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" event={"ID":"258a4e61-6fdc-495b-9f6d-fffb6e3ee702","Type":"ContainerStarted","Data":"d168be7079a1f1106fec98e6c7d4e88f66b04e8e23f45ab7f2a4c7a86ba6890a"} Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.511357 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovn-acl-logging/0.log" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.514371 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pqtt_6de8ddf6-38db-4394-8660-7e9659bf2bd2/ovn-controller/0.log" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.515785 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" event={"ID":"6de8ddf6-38db-4394-8660-7e9659bf2bd2","Type":"ContainerDied","Data":"aac58667e86956813631c75488951a17c47e7d516188cc5e5ee747d28143da9b"} Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.515921 4945 scope.go:117] "RemoveContainer" containerID="b7e9738cf3f80c415296bbc75a8735490f3d6112c9285ca52958a2fc67bda65e" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.516261 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7pqtt" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.551153 4945 scope.go:117] "RemoveContainer" containerID="228e4418ccd578fd9ad6b66c0c326cb3041cab2f6e57aee4a6fb5492759f47c0" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.575488 4945 scope.go:117] "RemoveContainer" containerID="251cfa64eaa8652ecd8ec701659c10a3acfce758a577824a239683699886c26b" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.597347 4945 scope.go:117] "RemoveContainer" containerID="3b3e7cc90e7bb37509d9aed57f72b3b3ebf9fd93dfb607d6f38538cce5c19a7d" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.628176 4945 scope.go:117] "RemoveContainer" containerID="8dc487bec75199e9accdedc5cf238c9b6e5b50cb9e2576fec93bfad6b2d62b59" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.630153 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7pqtt"] Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.637503 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7pqtt"] Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.645408 4945 scope.go:117] "RemoveContainer" containerID="85b4c68f67a265b42a7c70430a850947d97e6a84e442be6944ef3b3eb8d48563" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.667539 4945 scope.go:117] "RemoveContainer" containerID="7a8c7f5b2cdb797efb140cfbc5afeed86cf52c7845dbbc56cfd48b037d1c8e04" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.685050 4945 scope.go:117] "RemoveContainer" containerID="e4b88a9b3f8b143745e70a28893348675d27522a7f5cda127a1814e9e35e7488" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.705096 4945 scope.go:117] "RemoveContainer" containerID="9334f927e267ac43ed0d36612540b6d10c0d17b3f53994ae518b2b709c037883" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.774427 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6de8ddf6-38db-4394-8660-7e9659bf2bd2" path="/var/lib/kubelet/pods/6de8ddf6-38db-4394-8660-7e9659bf2bd2/volumes" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.795928 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.795972 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.796018 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.796522 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8f32c2d032f723cbbb763494c9ff11fd408a6db0a40ec4819ef209acfe780ed8"} pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 15:25:16 crc kubenswrapper[4945]: I1014 15:25:16.796575 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" containerID="cri-o://8f32c2d032f723cbbb763494c9ff11fd408a6db0a40ec4819ef209acfe780ed8" gracePeriod=600 Oct 14 15:25:17 crc kubenswrapper[4945]: I1014 15:25:17.523155 4945 generic.go:334] "Generic (PLEG): container finished" podID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerID="8f32c2d032f723cbbb763494c9ff11fd408a6db0a40ec4819ef209acfe780ed8" exitCode=0 Oct 14 15:25:17 crc kubenswrapper[4945]: I1014 15:25:17.523206 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerDied","Data":"8f32c2d032f723cbbb763494c9ff11fd408a6db0a40ec4819ef209acfe780ed8"} Oct 14 15:25:17 crc kubenswrapper[4945]: I1014 15:25:17.523710 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerStarted","Data":"1d2f9a019d15034c004b5841d70bbbf8261549fb60b58e5a8e606324eb14a78f"} Oct 14 15:25:17 crc kubenswrapper[4945]: I1014 15:25:17.523736 4945 scope.go:117] "RemoveContainer" containerID="c8c34e87ab2d9cc333cc18624592bb9519696247cdd2c0458ada848829da5a55" Oct 14 15:25:17 crc kubenswrapper[4945]: I1014 15:25:17.528311 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" event={"ID":"258a4e61-6fdc-495b-9f6d-fffb6e3ee702","Type":"ContainerStarted","Data":"828f281cef1aee15d66e9aca2fdcaaf4684615ed32a7fc6648af6ecf977c061f"} Oct 14 15:25:17 crc kubenswrapper[4945]: I1014 15:25:17.528352 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" event={"ID":"258a4e61-6fdc-495b-9f6d-fffb6e3ee702","Type":"ContainerStarted","Data":"e4b32af3747b79faea81fe4c32ec0b7a4ca461c024ff9595ff84d3967565a020"} Oct 14 15:25:17 crc kubenswrapper[4945]: I1014 15:25:17.528366 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" event={"ID":"258a4e61-6fdc-495b-9f6d-fffb6e3ee702","Type":"ContainerStarted","Data":"2888baa8df9fbad126ec3b02c2ee4704dca86a13b9379ea2916082acd01eaeeb"} Oct 14 15:25:17 crc kubenswrapper[4945]: I1014 15:25:17.528381 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" event={"ID":"258a4e61-6fdc-495b-9f6d-fffb6e3ee702","Type":"ContainerStarted","Data":"be511eb673ad1c5a19a0089cea62d3a540632b41ed097d13d424465e89f73a7b"} Oct 14 15:25:17 crc kubenswrapper[4945]: I1014 15:25:17.528392 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" event={"ID":"258a4e61-6fdc-495b-9f6d-fffb6e3ee702","Type":"ContainerStarted","Data":"4ff4189f21e20579202c9175852b9291f1753759de142c5f2c98b6a7faf79419"} Oct 14 15:25:17 crc kubenswrapper[4945]: I1014 15:25:17.528403 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" event={"ID":"258a4e61-6fdc-495b-9f6d-fffb6e3ee702","Type":"ContainerStarted","Data":"43d2cacd7c62d27e6ff4937e28f3ffa0427bb9af1899e2f6621172189a0cd621"} Oct 14 15:25:19 crc kubenswrapper[4945]: I1014 15:25:19.550308 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" event={"ID":"258a4e61-6fdc-495b-9f6d-fffb6e3ee702","Type":"ContainerStarted","Data":"c7c9d1b00b02e0d00bfab527a7e45cea9ba38686dd3df9bd84e3632a586e7aa7"} Oct 14 15:25:22 crc kubenswrapper[4945]: I1014 15:25:22.572903 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" event={"ID":"258a4e61-6fdc-495b-9f6d-fffb6e3ee702","Type":"ContainerStarted","Data":"1b558f80aaeb6389aa6e0be30dc9a071e7bec323dd898c29f0d0adb00ec31fd9"} Oct 14 15:25:22 crc kubenswrapper[4945]: I1014 15:25:22.573313 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:22 crc kubenswrapper[4945]: I1014 15:25:22.573325 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:22 crc kubenswrapper[4945]: I1014 15:25:22.573449 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:22 crc kubenswrapper[4945]: I1014 15:25:22.600430 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" podStartSLOduration=7.600411383 podStartE2EDuration="7.600411383s" podCreationTimestamp="2025-10-14 15:25:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:25:22.599727833 +0000 UTC m=+652.583776201" watchObservedRunningTime="2025-10-14 15:25:22.600411383 +0000 UTC m=+652.584459751" Oct 14 15:25:22 crc kubenswrapper[4945]: I1014 15:25:22.603807 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:22 crc kubenswrapper[4945]: I1014 15:25:22.604486 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:28 crc kubenswrapper[4945]: I1014 15:25:28.764061 4945 scope.go:117] "RemoveContainer" containerID="6daedef21246bc41a36b48797081a15d5a44d4a71e4e65c78f31574913dcdb57" Oct 14 15:25:28 crc kubenswrapper[4945]: E1014 15:25:28.764548 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-b76xs_openshift-multus(f82b9cd9-7a14-4dd9-bb0e-fde942389666)\"" pod="openshift-multus/multus-b76xs" podUID="f82b9cd9-7a14-4dd9-bb0e-fde942389666" Oct 14 15:25:40 crc kubenswrapper[4945]: I1014 15:25:40.765731 4945 scope.go:117] "RemoveContainer" containerID="6daedef21246bc41a36b48797081a15d5a44d4a71e4e65c78f31574913dcdb57" Oct 14 15:25:41 crc kubenswrapper[4945]: I1014 15:25:41.682889 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b76xs_f82b9cd9-7a14-4dd9-bb0e-fde942389666/kube-multus/2.log" Oct 14 15:25:41 crc kubenswrapper[4945]: I1014 15:25:41.683192 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b76xs" event={"ID":"f82b9cd9-7a14-4dd9-bb0e-fde942389666","Type":"ContainerStarted","Data":"5d719f43a315ee1e80592f8c0ea572f66892e4add6a596899c94c346c8855045"} Oct 14 15:25:46 crc kubenswrapper[4945]: I1014 15:25:46.087027 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6l8mm" Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.586985 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c"] Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.589149 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.591288 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.598812 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c"] Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.679552 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c\" (UID: \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.679661 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj5bl\" (UniqueName: \"kubernetes.io/projected/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-kube-api-access-mj5bl\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c\" (UID: \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.679732 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c\" (UID: \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.780164 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c\" (UID: \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.780258 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c\" (UID: \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.780399 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj5bl\" (UniqueName: \"kubernetes.io/projected/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-kube-api-access-mj5bl\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c\" (UID: \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.781027 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c\" (UID: \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.782099 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c\" (UID: \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.806760 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj5bl\" (UniqueName: \"kubernetes.io/projected/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-kube-api-access-mj5bl\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c\" (UID: \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:48 crc kubenswrapper[4945]: I1014 15:25:48.927167 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:49 crc kubenswrapper[4945]: I1014 15:25:49.123174 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c"] Oct 14 15:25:49 crc kubenswrapper[4945]: I1014 15:25:49.728853 4945 generic.go:334] "Generic (PLEG): container finished" podID="a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" containerID="aa66c5388474c28e82ba6162591382a9e295e7bf889f73695aae95194cbc2df0" exitCode=0 Oct 14 15:25:49 crc kubenswrapper[4945]: I1014 15:25:49.728958 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" event={"ID":"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398","Type":"ContainerDied","Data":"aa66c5388474c28e82ba6162591382a9e295e7bf889f73695aae95194cbc2df0"} Oct 14 15:25:49 crc kubenswrapper[4945]: I1014 15:25:49.728998 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" event={"ID":"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398","Type":"ContainerStarted","Data":"f1c78274a8a93ae8836276584ff303c88ee313930bf5f817f7e410e39b4568b4"} Oct 14 15:25:51 crc kubenswrapper[4945]: I1014 15:25:51.740593 4945 generic.go:334] "Generic (PLEG): container finished" podID="a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" containerID="b19b9c9b097940a8fc541f2dd201e6e21d1a30467c1c0a2fc4b96ad2cc61f481" exitCode=0 Oct 14 15:25:51 crc kubenswrapper[4945]: I1014 15:25:51.740692 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" event={"ID":"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398","Type":"ContainerDied","Data":"b19b9c9b097940a8fc541f2dd201e6e21d1a30467c1c0a2fc4b96ad2cc61f481"} Oct 14 15:25:52 crc kubenswrapper[4945]: I1014 15:25:52.749992 4945 generic.go:334] "Generic (PLEG): container finished" podID="a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" containerID="2a98f6354c5c33b07ca2c0ace8abed36306a2b83e3beb0b40634b62e7fc5010d" exitCode=0 Oct 14 15:25:52 crc kubenswrapper[4945]: I1014 15:25:52.750069 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" event={"ID":"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398","Type":"ContainerDied","Data":"2a98f6354c5c33b07ca2c0ace8abed36306a2b83e3beb0b40634b62e7fc5010d"} Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.042166 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.151998 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj5bl\" (UniqueName: \"kubernetes.io/projected/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-kube-api-access-mj5bl\") pod \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\" (UID: \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\") " Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.152073 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-bundle\") pod \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\" (UID: \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\") " Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.152192 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-util\") pod \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\" (UID: \"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398\") " Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.153752 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-bundle" (OuterVolumeSpecName: "bundle") pod "a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" (UID: "a90d4b1f-7443-474c-b4b7-7cb7ac9ef398"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.158767 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-kube-api-access-mj5bl" (OuterVolumeSpecName: "kube-api-access-mj5bl") pod "a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" (UID: "a90d4b1f-7443-474c-b4b7-7cb7ac9ef398"). InnerVolumeSpecName "kube-api-access-mj5bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.172245 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-util" (OuterVolumeSpecName: "util") pod "a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" (UID: "a90d4b1f-7443-474c-b4b7-7cb7ac9ef398"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.254439 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj5bl\" (UniqueName: \"kubernetes.io/projected/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-kube-api-access-mj5bl\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.254516 4945 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.254536 4945 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a90d4b1f-7443-474c-b4b7-7cb7ac9ef398-util\") on node \"crc\" DevicePath \"\"" Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.766136 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.769348 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c" event={"ID":"a90d4b1f-7443-474c-b4b7-7cb7ac9ef398","Type":"ContainerDied","Data":"f1c78274a8a93ae8836276584ff303c88ee313930bf5f817f7e410e39b4568b4"} Oct 14 15:25:54 crc kubenswrapper[4945]: I1014 15:25:54.769440 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1c78274a8a93ae8836276584ff303c88ee313930bf5f817f7e410e39b4568b4" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.220950 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-pz6t2"] Oct 14 15:25:56 crc kubenswrapper[4945]: E1014 15:25:56.221497 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" containerName="pull" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.221511 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" containerName="pull" Oct 14 15:25:56 crc kubenswrapper[4945]: E1014 15:25:56.221531 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" containerName="extract" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.221538 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" containerName="extract" Oct 14 15:25:56 crc kubenswrapper[4945]: E1014 15:25:56.221554 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" containerName="util" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.221561 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" containerName="util" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.221667 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a90d4b1f-7443-474c-b4b7-7cb7ac9ef398" containerName="extract" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.222183 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-pz6t2" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.224720 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.225421 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-vhfj2" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.228951 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.239840 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-pz6t2"] Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.279895 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgjgs\" (UniqueName: \"kubernetes.io/projected/27598c0b-8118-406d-b0dd-9e47a23343ad-kube-api-access-lgjgs\") pod \"nmstate-operator-858ddd8f98-pz6t2\" (UID: \"27598c0b-8118-406d-b0dd-9e47a23343ad\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-pz6t2" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.380467 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgjgs\" (UniqueName: \"kubernetes.io/projected/27598c0b-8118-406d-b0dd-9e47a23343ad-kube-api-access-lgjgs\") pod \"nmstate-operator-858ddd8f98-pz6t2\" (UID: \"27598c0b-8118-406d-b0dd-9e47a23343ad\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-pz6t2" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.398038 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgjgs\" (UniqueName: \"kubernetes.io/projected/27598c0b-8118-406d-b0dd-9e47a23343ad-kube-api-access-lgjgs\") pod \"nmstate-operator-858ddd8f98-pz6t2\" (UID: \"27598c0b-8118-406d-b0dd-9e47a23343ad\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-pz6t2" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.536610 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-pz6t2" Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.728757 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-pz6t2"] Oct 14 15:25:56 crc kubenswrapper[4945]: W1014 15:25:56.755528 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27598c0b_8118_406d_b0dd_9e47a23343ad.slice/crio-db5aabe6d4991482a8d5f883aa64af7f00b8e706709ad42de85a571f8147d1cd WatchSource:0}: Error finding container db5aabe6d4991482a8d5f883aa64af7f00b8e706709ad42de85a571f8147d1cd: Status 404 returned error can't find the container with id db5aabe6d4991482a8d5f883aa64af7f00b8e706709ad42de85a571f8147d1cd Oct 14 15:25:56 crc kubenswrapper[4945]: I1014 15:25:56.776692 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-pz6t2" event={"ID":"27598c0b-8118-406d-b0dd-9e47a23343ad","Type":"ContainerStarted","Data":"db5aabe6d4991482a8d5f883aa64af7f00b8e706709ad42de85a571f8147d1cd"} Oct 14 15:25:59 crc kubenswrapper[4945]: I1014 15:25:59.792988 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-pz6t2" event={"ID":"27598c0b-8118-406d-b0dd-9e47a23343ad","Type":"ContainerStarted","Data":"a4793cd6eeef0392530040c02e82260398927a970d87235349905629f15f969c"} Oct 14 15:25:59 crc kubenswrapper[4945]: I1014 15:25:59.812465 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-pz6t2" podStartSLOduration=0.926415213 podStartE2EDuration="3.812445108s" podCreationTimestamp="2025-10-14 15:25:56 +0000 UTC" firstStartedPulling="2025-10-14 15:25:56.757799649 +0000 UTC m=+686.741848027" lastFinishedPulling="2025-10-14 15:25:59.643829554 +0000 UTC m=+689.627877922" observedRunningTime="2025-10-14 15:25:59.808470772 +0000 UTC m=+689.792519150" watchObservedRunningTime="2025-10-14 15:25:59.812445108 +0000 UTC m=+689.796493476" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.717068 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-dh8w2"] Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.718844 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dh8w2" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.720394 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf"] Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.721198 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.721916 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-mjm6l" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.723072 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.727105 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/76343e33-0b0d-4df5-98cf-4d58247d41f7-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-p64sf\" (UID: \"76343e33-0b0d-4df5-98cf-4d58247d41f7\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.727234 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rbsl\" (UniqueName: \"kubernetes.io/projected/76343e33-0b0d-4df5-98cf-4d58247d41f7-kube-api-access-9rbsl\") pod \"nmstate-webhook-6cdbc54649-p64sf\" (UID: \"76343e33-0b0d-4df5-98cf-4d58247d41f7\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.727633 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xx8c\" (UniqueName: \"kubernetes.io/projected/97cf21fd-90e2-4591-8563-66afb4145fa4-kube-api-access-9xx8c\") pod \"nmstate-metrics-fdff9cb8d-dh8w2\" (UID: \"97cf21fd-90e2-4591-8563-66afb4145fa4\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dh8w2" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.740773 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf"] Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.753508 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-nx8z6"] Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.754145 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.796985 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-dh8w2"] Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.828183 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rbsl\" (UniqueName: \"kubernetes.io/projected/76343e33-0b0d-4df5-98cf-4d58247d41f7-kube-api-access-9rbsl\") pod \"nmstate-webhook-6cdbc54649-p64sf\" (UID: \"76343e33-0b0d-4df5-98cf-4d58247d41f7\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.828244 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/93cf1921-8abd-42e0-8f80-aa5023f63c96-nmstate-lock\") pod \"nmstate-handler-nx8z6\" (UID: \"93cf1921-8abd-42e0-8f80-aa5023f63c96\") " pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.828270 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d9n4\" (UniqueName: \"kubernetes.io/projected/93cf1921-8abd-42e0-8f80-aa5023f63c96-kube-api-access-6d9n4\") pod \"nmstate-handler-nx8z6\" (UID: \"93cf1921-8abd-42e0-8f80-aa5023f63c96\") " pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.828296 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/93cf1921-8abd-42e0-8f80-aa5023f63c96-ovs-socket\") pod \"nmstate-handler-nx8z6\" (UID: \"93cf1921-8abd-42e0-8f80-aa5023f63c96\") " pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.828318 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/93cf1921-8abd-42e0-8f80-aa5023f63c96-dbus-socket\") pod \"nmstate-handler-nx8z6\" (UID: \"93cf1921-8abd-42e0-8f80-aa5023f63c96\") " pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.828371 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xx8c\" (UniqueName: \"kubernetes.io/projected/97cf21fd-90e2-4591-8563-66afb4145fa4-kube-api-access-9xx8c\") pod \"nmstate-metrics-fdff9cb8d-dh8w2\" (UID: \"97cf21fd-90e2-4591-8563-66afb4145fa4\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dh8w2" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.828413 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/76343e33-0b0d-4df5-98cf-4d58247d41f7-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-p64sf\" (UID: \"76343e33-0b0d-4df5-98cf-4d58247d41f7\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" Oct 14 15:26:00 crc kubenswrapper[4945]: E1014 15:26:00.828502 4945 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 14 15:26:00 crc kubenswrapper[4945]: E1014 15:26:00.828569 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/76343e33-0b0d-4df5-98cf-4d58247d41f7-tls-key-pair podName:76343e33-0b0d-4df5-98cf-4d58247d41f7 nodeName:}" failed. No retries permitted until 2025-10-14 15:26:01.328532489 +0000 UTC m=+691.312580857 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/76343e33-0b0d-4df5-98cf-4d58247d41f7-tls-key-pair") pod "nmstate-webhook-6cdbc54649-p64sf" (UID: "76343e33-0b0d-4df5-98cf-4d58247d41f7") : secret "openshift-nmstate-webhook" not found Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.843989 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l"] Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.844742 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.849546 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-9wwq2" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.849704 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.849807 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.855625 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l"] Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.856054 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rbsl\" (UniqueName: \"kubernetes.io/projected/76343e33-0b0d-4df5-98cf-4d58247d41f7-kube-api-access-9rbsl\") pod \"nmstate-webhook-6cdbc54649-p64sf\" (UID: \"76343e33-0b0d-4df5-98cf-4d58247d41f7\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.856053 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xx8c\" (UniqueName: \"kubernetes.io/projected/97cf21fd-90e2-4591-8563-66afb4145fa4-kube-api-access-9xx8c\") pod \"nmstate-metrics-fdff9cb8d-dh8w2\" (UID: \"97cf21fd-90e2-4591-8563-66afb4145fa4\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dh8w2" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.929402 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/93cf1921-8abd-42e0-8f80-aa5023f63c96-nmstate-lock\") pod \"nmstate-handler-nx8z6\" (UID: \"93cf1921-8abd-42e0-8f80-aa5023f63c96\") " pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.929436 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d9n4\" (UniqueName: \"kubernetes.io/projected/93cf1921-8abd-42e0-8f80-aa5023f63c96-kube-api-access-6d9n4\") pod \"nmstate-handler-nx8z6\" (UID: \"93cf1921-8abd-42e0-8f80-aa5023f63c96\") " pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.929451 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/93cf1921-8abd-42e0-8f80-aa5023f63c96-ovs-socket\") pod \"nmstate-handler-nx8z6\" (UID: \"93cf1921-8abd-42e0-8f80-aa5023f63c96\") " pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.929470 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxbhn\" (UniqueName: \"kubernetes.io/projected/f481cdcc-3f50-4aba-a28a-500e4b19bcff-kube-api-access-jxbhn\") pod \"nmstate-console-plugin-6b874cbd85-27p5l\" (UID: \"f481cdcc-3f50-4aba-a28a-500e4b19bcff\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.929495 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/93cf1921-8abd-42e0-8f80-aa5023f63c96-dbus-socket\") pod \"nmstate-handler-nx8z6\" (UID: \"93cf1921-8abd-42e0-8f80-aa5023f63c96\") " pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.929545 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/93cf1921-8abd-42e0-8f80-aa5023f63c96-ovs-socket\") pod \"nmstate-handler-nx8z6\" (UID: \"93cf1921-8abd-42e0-8f80-aa5023f63c96\") " pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.929557 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f481cdcc-3f50-4aba-a28a-500e4b19bcff-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-27p5l\" (UID: \"f481cdcc-3f50-4aba-a28a-500e4b19bcff\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.929601 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f481cdcc-3f50-4aba-a28a-500e4b19bcff-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-27p5l\" (UID: \"f481cdcc-3f50-4aba-a28a-500e4b19bcff\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.929747 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/93cf1921-8abd-42e0-8f80-aa5023f63c96-dbus-socket\") pod \"nmstate-handler-nx8z6\" (UID: \"93cf1921-8abd-42e0-8f80-aa5023f63c96\") " pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.929794 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/93cf1921-8abd-42e0-8f80-aa5023f63c96-nmstate-lock\") pod \"nmstate-handler-nx8z6\" (UID: \"93cf1921-8abd-42e0-8f80-aa5023f63c96\") " pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:00 crc kubenswrapper[4945]: I1014 15:26:00.948899 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d9n4\" (UniqueName: \"kubernetes.io/projected/93cf1921-8abd-42e0-8f80-aa5023f63c96-kube-api-access-6d9n4\") pod \"nmstate-handler-nx8z6\" (UID: \"93cf1921-8abd-42e0-8f80-aa5023f63c96\") " pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.030424 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f481cdcc-3f50-4aba-a28a-500e4b19bcff-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-27p5l\" (UID: \"f481cdcc-3f50-4aba-a28a-500e4b19bcff\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.030485 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f481cdcc-3f50-4aba-a28a-500e4b19bcff-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-27p5l\" (UID: \"f481cdcc-3f50-4aba-a28a-500e4b19bcff\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.030540 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxbhn\" (UniqueName: \"kubernetes.io/projected/f481cdcc-3f50-4aba-a28a-500e4b19bcff-kube-api-access-jxbhn\") pod \"nmstate-console-plugin-6b874cbd85-27p5l\" (UID: \"f481cdcc-3f50-4aba-a28a-500e4b19bcff\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.032773 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f481cdcc-3f50-4aba-a28a-500e4b19bcff-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-27p5l\" (UID: \"f481cdcc-3f50-4aba-a28a-500e4b19bcff\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.036612 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f481cdcc-3f50-4aba-a28a-500e4b19bcff-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-27p5l\" (UID: \"f481cdcc-3f50-4aba-a28a-500e4b19bcff\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.043302 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dh8w2" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.066175 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6dfdbf4bb9-5mfdl"] Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.066948 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.068634 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.071379 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxbhn\" (UniqueName: \"kubernetes.io/projected/f481cdcc-3f50-4aba-a28a-500e4b19bcff-kube-api-access-jxbhn\") pod \"nmstate-console-plugin-6b874cbd85-27p5l\" (UID: \"f481cdcc-3f50-4aba-a28a-500e4b19bcff\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.082415 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6dfdbf4bb9-5mfdl"] Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.196571 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.233912 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-console-oauth-config\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.233968 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-service-ca\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.234154 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-console-config\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.234234 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-trusted-ca-bundle\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.234307 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lnm7\" (UniqueName: \"kubernetes.io/projected/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-kube-api-access-9lnm7\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.234395 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-console-serving-cert\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.234457 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-oauth-serving-cert\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.261888 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-dh8w2"] Oct 14 15:26:01 crc kubenswrapper[4945]: W1014 15:26:01.265296 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97cf21fd_90e2_4591_8563_66afb4145fa4.slice/crio-8a769b7396f86fcd2e35c3ffd85c5bd391dc706d7ceb0b77b6caa1925812c21c WatchSource:0}: Error finding container 8a769b7396f86fcd2e35c3ffd85c5bd391dc706d7ceb0b77b6caa1925812c21c: Status 404 returned error can't find the container with id 8a769b7396f86fcd2e35c3ffd85c5bd391dc706d7ceb0b77b6caa1925812c21c Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.336083 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/76343e33-0b0d-4df5-98cf-4d58247d41f7-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-p64sf\" (UID: \"76343e33-0b0d-4df5-98cf-4d58247d41f7\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.336631 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-console-config\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.336662 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-trusted-ca-bundle\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.336749 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lnm7\" (UniqueName: \"kubernetes.io/projected/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-kube-api-access-9lnm7\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.337318 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-console-serving-cert\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.337390 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-oauth-serving-cert\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.337789 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-trusted-ca-bundle\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.338344 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-console-config\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.338512 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-oauth-serving-cert\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.338616 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-console-oauth-config\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.339182 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-service-ca\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.340186 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-service-ca\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.343626 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/76343e33-0b0d-4df5-98cf-4d58247d41f7-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-p64sf\" (UID: \"76343e33-0b0d-4df5-98cf-4d58247d41f7\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.343829 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-console-oauth-config\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.343850 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-console-serving-cert\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.353053 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.355785 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lnm7\" (UniqueName: \"kubernetes.io/projected/ea3baea3-92bb-41ba-b80a-cf0f9fd04cff-kube-api-access-9lnm7\") pod \"console-6dfdbf4bb9-5mfdl\" (UID: \"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff\") " pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.393936 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l"] Oct 14 15:26:01 crc kubenswrapper[4945]: W1014 15:26:01.403262 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf481cdcc_3f50_4aba_a28a_500e4b19bcff.slice/crio-ed424f8156a4ac39ebb499d79770140ada8516be4c1bc35d05bf9fc8624a9581 WatchSource:0}: Error finding container ed424f8156a4ac39ebb499d79770140ada8516be4c1bc35d05bf9fc8624a9581: Status 404 returned error can't find the container with id ed424f8156a4ac39ebb499d79770140ada8516be4c1bc35d05bf9fc8624a9581 Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.404563 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.539905 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf"] Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.813353 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" event={"ID":"f481cdcc-3f50-4aba-a28a-500e4b19bcff","Type":"ContainerStarted","Data":"ed424f8156a4ac39ebb499d79770140ada8516be4c1bc35d05bf9fc8624a9581"} Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.814252 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dh8w2" event={"ID":"97cf21fd-90e2-4591-8563-66afb4145fa4","Type":"ContainerStarted","Data":"8a769b7396f86fcd2e35c3ffd85c5bd391dc706d7ceb0b77b6caa1925812c21c"} Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.814744 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6dfdbf4bb9-5mfdl"] Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.815688 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" event={"ID":"76343e33-0b0d-4df5-98cf-4d58247d41f7","Type":"ContainerStarted","Data":"75abb48bf4c76f3cb5ccb0ccc18799a53914338f584cf9be36f5d4b49cdec696"} Oct 14 15:26:01 crc kubenswrapper[4945]: I1014 15:26:01.816842 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-nx8z6" event={"ID":"93cf1921-8abd-42e0-8f80-aa5023f63c96","Type":"ContainerStarted","Data":"6537e3362b885991d758a2658690415c30f71eb0b4b39c46d1d597942bdf2d4a"} Oct 14 15:26:01 crc kubenswrapper[4945]: W1014 15:26:01.821879 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3baea3_92bb_41ba_b80a_cf0f9fd04cff.slice/crio-a56b2f72ebb72e83ab7ea65d26d80914b6971ba46c1e0cc102bc3ee9aad8517b WatchSource:0}: Error finding container a56b2f72ebb72e83ab7ea65d26d80914b6971ba46c1e0cc102bc3ee9aad8517b: Status 404 returned error can't find the container with id a56b2f72ebb72e83ab7ea65d26d80914b6971ba46c1e0cc102bc3ee9aad8517b Oct 14 15:26:02 crc kubenswrapper[4945]: I1014 15:26:02.824277 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6dfdbf4bb9-5mfdl" event={"ID":"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff","Type":"ContainerStarted","Data":"37e16d653a3f4d770eb26e83a650b64a332d5aa9c8965f6730784b32d912838a"} Oct 14 15:26:02 crc kubenswrapper[4945]: I1014 15:26:02.824650 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6dfdbf4bb9-5mfdl" event={"ID":"ea3baea3-92bb-41ba-b80a-cf0f9fd04cff","Type":"ContainerStarted","Data":"a56b2f72ebb72e83ab7ea65d26d80914b6971ba46c1e0cc102bc3ee9aad8517b"} Oct 14 15:26:02 crc kubenswrapper[4945]: I1014 15:26:02.842512 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6dfdbf4bb9-5mfdl" podStartSLOduration=1.84248995 podStartE2EDuration="1.84248995s" podCreationTimestamp="2025-10-14 15:26:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:26:02.83939244 +0000 UTC m=+692.823440818" watchObservedRunningTime="2025-10-14 15:26:02.84248995 +0000 UTC m=+692.826538318" Oct 14 15:26:05 crc kubenswrapper[4945]: I1014 15:26:05.855312 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dh8w2" event={"ID":"97cf21fd-90e2-4591-8563-66afb4145fa4","Type":"ContainerStarted","Data":"e8b240c43ac1d60b0661e775ecba1c99c9792d505db92b7eb7592de75c029ec0"} Oct 14 15:26:05 crc kubenswrapper[4945]: I1014 15:26:05.857321 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" event={"ID":"76343e33-0b0d-4df5-98cf-4d58247d41f7","Type":"ContainerStarted","Data":"2840a0a28243552c1f7311999f43e30ccaf1abc4940f9a1a0309573399a9f568"} Oct 14 15:26:05 crc kubenswrapper[4945]: I1014 15:26:05.857461 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" Oct 14 15:26:05 crc kubenswrapper[4945]: I1014 15:26:05.858789 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-nx8z6" event={"ID":"93cf1921-8abd-42e0-8f80-aa5023f63c96","Type":"ContainerStarted","Data":"8e54fe2a020c636ce6728257366b4842df084f93c93d095b828a9a1d85680efc"} Oct 14 15:26:05 crc kubenswrapper[4945]: I1014 15:26:05.858909 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:05 crc kubenswrapper[4945]: I1014 15:26:05.860784 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" event={"ID":"f481cdcc-3f50-4aba-a28a-500e4b19bcff","Type":"ContainerStarted","Data":"40c6eb22e15ce53519927a8aea5fda45fb2ccbcd3ba86df29611952292423c40"} Oct 14 15:26:05 crc kubenswrapper[4945]: I1014 15:26:05.883145 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" podStartSLOduration=2.099278501 podStartE2EDuration="5.883123221s" podCreationTimestamp="2025-10-14 15:26:00 +0000 UTC" firstStartedPulling="2025-10-14 15:26:01.541313201 +0000 UTC m=+691.525361569" lastFinishedPulling="2025-10-14 15:26:05.325157921 +0000 UTC m=+695.309206289" observedRunningTime="2025-10-14 15:26:05.876844678 +0000 UTC m=+695.860893066" watchObservedRunningTime="2025-10-14 15:26:05.883123221 +0000 UTC m=+695.867171589" Oct 14 15:26:05 crc kubenswrapper[4945]: I1014 15:26:05.895014 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-27p5l" podStartSLOduration=1.9866382200000001 podStartE2EDuration="5.894991407s" podCreationTimestamp="2025-10-14 15:26:00 +0000 UTC" firstStartedPulling="2025-10-14 15:26:01.405508264 +0000 UTC m=+691.389556632" lastFinishedPulling="2025-10-14 15:26:05.313861461 +0000 UTC m=+695.297909819" observedRunningTime="2025-10-14 15:26:05.892839454 +0000 UTC m=+695.876887822" watchObservedRunningTime="2025-10-14 15:26:05.894991407 +0000 UTC m=+695.879039775" Oct 14 15:26:05 crc kubenswrapper[4945]: I1014 15:26:05.917987 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-nx8z6" podStartSLOduration=1.633406934 podStartE2EDuration="5.917967976s" podCreationTimestamp="2025-10-14 15:26:00 +0000 UTC" firstStartedPulling="2025-10-14 15:26:01.10720327 +0000 UTC m=+691.091251638" lastFinishedPulling="2025-10-14 15:26:05.391764312 +0000 UTC m=+695.375812680" observedRunningTime="2025-10-14 15:26:05.915261458 +0000 UTC m=+695.899309826" watchObservedRunningTime="2025-10-14 15:26:05.917967976 +0000 UTC m=+695.902016364" Oct 14 15:26:08 crc kubenswrapper[4945]: I1014 15:26:08.879164 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dh8w2" event={"ID":"97cf21fd-90e2-4591-8563-66afb4145fa4","Type":"ContainerStarted","Data":"99cf1300c9cac3b612fce25a390911a2ef186874deb660977e471a46f221beda"} Oct 14 15:26:08 crc kubenswrapper[4945]: I1014 15:26:08.893644 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dh8w2" podStartSLOduration=1.5376877759999998 podStartE2EDuration="8.893625293s" podCreationTimestamp="2025-10-14 15:26:00 +0000 UTC" firstStartedPulling="2025-10-14 15:26:01.269597103 +0000 UTC m=+691.253645471" lastFinishedPulling="2025-10-14 15:26:08.62553461 +0000 UTC m=+698.609582988" observedRunningTime="2025-10-14 15:26:08.891349227 +0000 UTC m=+698.875397595" watchObservedRunningTime="2025-10-14 15:26:08.893625293 +0000 UTC m=+698.877673661" Oct 14 15:26:11 crc kubenswrapper[4945]: I1014 15:26:11.101934 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-nx8z6" Oct 14 15:26:11 crc kubenswrapper[4945]: I1014 15:26:11.405130 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:11 crc kubenswrapper[4945]: I1014 15:26:11.405482 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:11 crc kubenswrapper[4945]: I1014 15:26:11.409423 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:11 crc kubenswrapper[4945]: I1014 15:26:11.905587 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6dfdbf4bb9-5mfdl" Oct 14 15:26:11 crc kubenswrapper[4945]: I1014 15:26:11.965798 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ccxsw"] Oct 14 15:26:21 crc kubenswrapper[4945]: I1014 15:26:21.366515 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-p64sf" Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.446761 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp"] Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.448525 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.450104 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.455521 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp"] Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.645855 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp\" (UID: \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.645943 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp\" (UID: \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.645986 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f2mn\" (UniqueName: \"kubernetes.io/projected/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-kube-api-access-8f2mn\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp\" (UID: \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.747575 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp\" (UID: \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.747625 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp\" (UID: \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.747669 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f2mn\" (UniqueName: \"kubernetes.io/projected/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-kube-api-access-8f2mn\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp\" (UID: \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.748254 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp\" (UID: \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.748254 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp\" (UID: \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:34 crc kubenswrapper[4945]: I1014 15:26:34.766605 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f2mn\" (UniqueName: \"kubernetes.io/projected/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-kube-api-access-8f2mn\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp\" (UID: \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:35 crc kubenswrapper[4945]: I1014 15:26:35.065146 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:35 crc kubenswrapper[4945]: I1014 15:26:35.315863 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp"] Oct 14 15:26:36 crc kubenswrapper[4945]: I1014 15:26:36.049813 4945 generic.go:334] "Generic (PLEG): container finished" podID="c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" containerID="ac738be6c2762398b3d7bf71624c186c9396d53a3d30a79a585c5d3ce7d99076" exitCode=0 Oct 14 15:26:36 crc kubenswrapper[4945]: I1014 15:26:36.049919 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" event={"ID":"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d","Type":"ContainerDied","Data":"ac738be6c2762398b3d7bf71624c186c9396d53a3d30a79a585c5d3ce7d99076"} Oct 14 15:26:36 crc kubenswrapper[4945]: I1014 15:26:36.050191 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" event={"ID":"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d","Type":"ContainerStarted","Data":"974813c4e092a0a9cf5a4d38f8d9ac15de53282414887e5d73e6af9eb3289b55"} Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.022995 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-ccxsw" podUID="06ecdccd-2560-48ea-a7f3-87e3e75488a0" containerName="console" containerID="cri-o://b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277" gracePeriod=15 Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.519371 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ccxsw_06ecdccd-2560-48ea-a7f3-87e3e75488a0/console/0.log" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.519666 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.696599 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-service-ca\") pod \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.696651 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-serving-cert\") pod \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.696677 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-trusted-ca-bundle\") pod \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.696739 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgjgp\" (UniqueName: \"kubernetes.io/projected/06ecdccd-2560-48ea-a7f3-87e3e75488a0-kube-api-access-mgjgp\") pod \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.696782 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-oauth-serving-cert\") pod \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.696820 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-config\") pod \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.696854 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-oauth-config\") pod \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\" (UID: \"06ecdccd-2560-48ea-a7f3-87e3e75488a0\") " Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.697783 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "06ecdccd-2560-48ea-a7f3-87e3e75488a0" (UID: "06ecdccd-2560-48ea-a7f3-87e3e75488a0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.697838 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-service-ca" (OuterVolumeSpecName: "service-ca") pod "06ecdccd-2560-48ea-a7f3-87e3e75488a0" (UID: "06ecdccd-2560-48ea-a7f3-87e3e75488a0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.697926 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "06ecdccd-2560-48ea-a7f3-87e3e75488a0" (UID: "06ecdccd-2560-48ea-a7f3-87e3e75488a0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.698021 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-config" (OuterVolumeSpecName: "console-config") pod "06ecdccd-2560-48ea-a7f3-87e3e75488a0" (UID: "06ecdccd-2560-48ea-a7f3-87e3e75488a0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.703183 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06ecdccd-2560-48ea-a7f3-87e3e75488a0-kube-api-access-mgjgp" (OuterVolumeSpecName: "kube-api-access-mgjgp") pod "06ecdccd-2560-48ea-a7f3-87e3e75488a0" (UID: "06ecdccd-2560-48ea-a7f3-87e3e75488a0"). InnerVolumeSpecName "kube-api-access-mgjgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.703814 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "06ecdccd-2560-48ea-a7f3-87e3e75488a0" (UID: "06ecdccd-2560-48ea-a7f3-87e3e75488a0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.704190 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "06ecdccd-2560-48ea-a7f3-87e3e75488a0" (UID: "06ecdccd-2560-48ea-a7f3-87e3e75488a0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.798299 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgjgp\" (UniqueName: \"kubernetes.io/projected/06ecdccd-2560-48ea-a7f3-87e3e75488a0-kube-api-access-mgjgp\") on node \"crc\" DevicePath \"\"" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.798335 4945 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.798350 4945 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.798361 4945 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.798372 4945 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.798383 4945 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06ecdccd-2560-48ea-a7f3-87e3e75488a0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:26:37 crc kubenswrapper[4945]: I1014 15:26:37.798395 4945 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06ecdccd-2560-48ea-a7f3-87e3e75488a0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:26:38 crc kubenswrapper[4945]: I1014 15:26:38.062319 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ccxsw_06ecdccd-2560-48ea-a7f3-87e3e75488a0/console/0.log" Oct 14 15:26:38 crc kubenswrapper[4945]: I1014 15:26:38.062366 4945 generic.go:334] "Generic (PLEG): container finished" podID="06ecdccd-2560-48ea-a7f3-87e3e75488a0" containerID="b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277" exitCode=2 Oct 14 15:26:38 crc kubenswrapper[4945]: I1014 15:26:38.062394 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ccxsw" event={"ID":"06ecdccd-2560-48ea-a7f3-87e3e75488a0","Type":"ContainerDied","Data":"b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277"} Oct 14 15:26:38 crc kubenswrapper[4945]: I1014 15:26:38.062419 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ccxsw" event={"ID":"06ecdccd-2560-48ea-a7f3-87e3e75488a0","Type":"ContainerDied","Data":"354b14026fcbff4bf145f5c84b4f72443811c15403dc6c5671c2b01813577872"} Oct 14 15:26:38 crc kubenswrapper[4945]: I1014 15:26:38.062441 4945 scope.go:117] "RemoveContainer" containerID="b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277" Oct 14 15:26:38 crc kubenswrapper[4945]: I1014 15:26:38.062457 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ccxsw" Oct 14 15:26:38 crc kubenswrapper[4945]: I1014 15:26:38.092844 4945 scope.go:117] "RemoveContainer" containerID="b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277" Oct 14 15:26:38 crc kubenswrapper[4945]: E1014 15:26:38.093241 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277\": container with ID starting with b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277 not found: ID does not exist" containerID="b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277" Oct 14 15:26:38 crc kubenswrapper[4945]: I1014 15:26:38.093285 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277"} err="failed to get container status \"b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277\": rpc error: code = NotFound desc = could not find container \"b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277\": container with ID starting with b08f4ec27ec27e65a7dc2dc0b0f3396a0966d74d9975015bcae0f12daf97a277 not found: ID does not exist" Oct 14 15:26:38 crc kubenswrapper[4945]: I1014 15:26:38.099810 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ccxsw"] Oct 14 15:26:38 crc kubenswrapper[4945]: I1014 15:26:38.106732 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-ccxsw"] Oct 14 15:26:38 crc kubenswrapper[4945]: I1014 15:26:38.773293 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06ecdccd-2560-48ea-a7f3-87e3e75488a0" path="/var/lib/kubelet/pods/06ecdccd-2560-48ea-a7f3-87e3e75488a0/volumes" Oct 14 15:26:39 crc kubenswrapper[4945]: I1014 15:26:39.075210 4945 generic.go:334] "Generic (PLEG): container finished" podID="c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" containerID="958301eacde7a3fe46113dd628973bf486c5efb8edb4bd03a354c279bd0d18a0" exitCode=0 Oct 14 15:26:39 crc kubenswrapper[4945]: I1014 15:26:39.075280 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" event={"ID":"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d","Type":"ContainerDied","Data":"958301eacde7a3fe46113dd628973bf486c5efb8edb4bd03a354c279bd0d18a0"} Oct 14 15:26:40 crc kubenswrapper[4945]: I1014 15:26:40.089466 4945 generic.go:334] "Generic (PLEG): container finished" podID="c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" containerID="7a8e7981f47f7e7f1415bf7bd9ee167bf4ac911316754a82f4c0a3fb0c0468b7" exitCode=0 Oct 14 15:26:40 crc kubenswrapper[4945]: I1014 15:26:40.089518 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" event={"ID":"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d","Type":"ContainerDied","Data":"7a8e7981f47f7e7f1415bf7bd9ee167bf4ac911316754a82f4c0a3fb0c0468b7"} Oct 14 15:26:41 crc kubenswrapper[4945]: I1014 15:26:41.343842 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:41 crc kubenswrapper[4945]: I1014 15:26:41.444549 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-util\") pod \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\" (UID: \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\") " Oct 14 15:26:41 crc kubenswrapper[4945]: I1014 15:26:41.444674 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-bundle\") pod \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\" (UID: \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\") " Oct 14 15:26:41 crc kubenswrapper[4945]: I1014 15:26:41.444736 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f2mn\" (UniqueName: \"kubernetes.io/projected/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-kube-api-access-8f2mn\") pod \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\" (UID: \"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d\") " Oct 14 15:26:41 crc kubenswrapper[4945]: I1014 15:26:41.445598 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-bundle" (OuterVolumeSpecName: "bundle") pod "c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" (UID: "c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:26:41 crc kubenswrapper[4945]: I1014 15:26:41.449525 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-kube-api-access-8f2mn" (OuterVolumeSpecName: "kube-api-access-8f2mn") pod "c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" (UID: "c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d"). InnerVolumeSpecName "kube-api-access-8f2mn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:26:41 crc kubenswrapper[4945]: I1014 15:26:41.484259 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-util" (OuterVolumeSpecName: "util") pod "c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" (UID: "c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:26:41 crc kubenswrapper[4945]: I1014 15:26:41.546175 4945 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:26:41 crc kubenswrapper[4945]: I1014 15:26:41.546218 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f2mn\" (UniqueName: \"kubernetes.io/projected/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-kube-api-access-8f2mn\") on node \"crc\" DevicePath \"\"" Oct 14 15:26:41 crc kubenswrapper[4945]: I1014 15:26:41.546233 4945 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d-util\") on node \"crc\" DevicePath \"\"" Oct 14 15:26:42 crc kubenswrapper[4945]: I1014 15:26:42.107665 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" event={"ID":"c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d","Type":"ContainerDied","Data":"974813c4e092a0a9cf5a4d38f8d9ac15de53282414887e5d73e6af9eb3289b55"} Oct 14 15:26:42 crc kubenswrapper[4945]: I1014 15:26:42.107711 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="974813c4e092a0a9cf5a4d38f8d9ac15de53282414887e5d73e6af9eb3289b55" Oct 14 15:26:42 crc kubenswrapper[4945]: I1014 15:26:42.107781 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.776025 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd"] Oct 14 15:26:49 crc kubenswrapper[4945]: E1014 15:26:49.776817 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06ecdccd-2560-48ea-a7f3-87e3e75488a0" containerName="console" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.776831 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="06ecdccd-2560-48ea-a7f3-87e3e75488a0" containerName="console" Oct 14 15:26:49 crc kubenswrapper[4945]: E1014 15:26:49.776845 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" containerName="pull" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.776854 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" containerName="pull" Oct 14 15:26:49 crc kubenswrapper[4945]: E1014 15:26:49.776864 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" containerName="util" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.776870 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" containerName="util" Oct 14 15:26:49 crc kubenswrapper[4945]: E1014 15:26:49.776926 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" containerName="extract" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.776934 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" containerName="extract" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.777074 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d" containerName="extract" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.777088 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="06ecdccd-2560-48ea-a7f3-87e3e75488a0" containerName="console" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.777574 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.780135 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-xnlkr" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.780210 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.780967 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.781492 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.797023 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.804433 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd"] Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.948761 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/740d5a8e-2e4a-4fd1-8412-d5b989734f2f-webhook-cert\") pod \"metallb-operator-controller-manager-9f47f9675-47zvd\" (UID: \"740d5a8e-2e4a-4fd1-8412-d5b989734f2f\") " pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.949062 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/740d5a8e-2e4a-4fd1-8412-d5b989734f2f-apiservice-cert\") pod \"metallb-operator-controller-manager-9f47f9675-47zvd\" (UID: \"740d5a8e-2e4a-4fd1-8412-d5b989734f2f\") " pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:26:49 crc kubenswrapper[4945]: I1014 15:26:49.949130 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7sjn\" (UniqueName: \"kubernetes.io/projected/740d5a8e-2e4a-4fd1-8412-d5b989734f2f-kube-api-access-c7sjn\") pod \"metallb-operator-controller-manager-9f47f9675-47zvd\" (UID: \"740d5a8e-2e4a-4fd1-8412-d5b989734f2f\") " pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.050366 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/740d5a8e-2e4a-4fd1-8412-d5b989734f2f-webhook-cert\") pod \"metallb-operator-controller-manager-9f47f9675-47zvd\" (UID: \"740d5a8e-2e4a-4fd1-8412-d5b989734f2f\") " pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.050482 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/740d5a8e-2e4a-4fd1-8412-d5b989734f2f-apiservice-cert\") pod \"metallb-operator-controller-manager-9f47f9675-47zvd\" (UID: \"740d5a8e-2e4a-4fd1-8412-d5b989734f2f\") " pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.050515 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7sjn\" (UniqueName: \"kubernetes.io/projected/740d5a8e-2e4a-4fd1-8412-d5b989734f2f-kube-api-access-c7sjn\") pod \"metallb-operator-controller-manager-9f47f9675-47zvd\" (UID: \"740d5a8e-2e4a-4fd1-8412-d5b989734f2f\") " pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.057674 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/740d5a8e-2e4a-4fd1-8412-d5b989734f2f-apiservice-cert\") pod \"metallb-operator-controller-manager-9f47f9675-47zvd\" (UID: \"740d5a8e-2e4a-4fd1-8412-d5b989734f2f\") " pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.070565 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/740d5a8e-2e4a-4fd1-8412-d5b989734f2f-webhook-cert\") pod \"metallb-operator-controller-manager-9f47f9675-47zvd\" (UID: \"740d5a8e-2e4a-4fd1-8412-d5b989734f2f\") " pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.077534 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7sjn\" (UniqueName: \"kubernetes.io/projected/740d5a8e-2e4a-4fd1-8412-d5b989734f2f-kube-api-access-c7sjn\") pod \"metallb-operator-controller-manager-9f47f9675-47zvd\" (UID: \"740d5a8e-2e4a-4fd1-8412-d5b989734f2f\") " pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.102306 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.237718 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w"] Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.238626 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.243398 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.243930 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-qfmdd" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.244062 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.257539 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w"] Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.356231 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/59c45798-dcb8-4fae-ba47-82a861d97023-webhook-cert\") pod \"metallb-operator-webhook-server-6545bcf4ff-f248w\" (UID: \"59c45798-dcb8-4fae-ba47-82a861d97023\") " pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.356553 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/59c45798-dcb8-4fae-ba47-82a861d97023-apiservice-cert\") pod \"metallb-operator-webhook-server-6545bcf4ff-f248w\" (UID: \"59c45798-dcb8-4fae-ba47-82a861d97023\") " pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.356589 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tjv5\" (UniqueName: \"kubernetes.io/projected/59c45798-dcb8-4fae-ba47-82a861d97023-kube-api-access-8tjv5\") pod \"metallb-operator-webhook-server-6545bcf4ff-f248w\" (UID: \"59c45798-dcb8-4fae-ba47-82a861d97023\") " pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.386058 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd"] Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.457617 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/59c45798-dcb8-4fae-ba47-82a861d97023-webhook-cert\") pod \"metallb-operator-webhook-server-6545bcf4ff-f248w\" (UID: \"59c45798-dcb8-4fae-ba47-82a861d97023\") " pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.457699 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/59c45798-dcb8-4fae-ba47-82a861d97023-apiservice-cert\") pod \"metallb-operator-webhook-server-6545bcf4ff-f248w\" (UID: \"59c45798-dcb8-4fae-ba47-82a861d97023\") " pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.457727 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tjv5\" (UniqueName: \"kubernetes.io/projected/59c45798-dcb8-4fae-ba47-82a861d97023-kube-api-access-8tjv5\") pod \"metallb-operator-webhook-server-6545bcf4ff-f248w\" (UID: \"59c45798-dcb8-4fae-ba47-82a861d97023\") " pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.462898 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/59c45798-dcb8-4fae-ba47-82a861d97023-webhook-cert\") pod \"metallb-operator-webhook-server-6545bcf4ff-f248w\" (UID: \"59c45798-dcb8-4fae-ba47-82a861d97023\") " pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.463011 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/59c45798-dcb8-4fae-ba47-82a861d97023-apiservice-cert\") pod \"metallb-operator-webhook-server-6545bcf4ff-f248w\" (UID: \"59c45798-dcb8-4fae-ba47-82a861d97023\") " pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.477550 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tjv5\" (UniqueName: \"kubernetes.io/projected/59c45798-dcb8-4fae-ba47-82a861d97023-kube-api-access-8tjv5\") pod \"metallb-operator-webhook-server-6545bcf4ff-f248w\" (UID: \"59c45798-dcb8-4fae-ba47-82a861d97023\") " pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:26:50 crc kubenswrapper[4945]: I1014 15:26:50.566364 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:26:51 crc kubenswrapper[4945]: I1014 15:26:51.102140 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w"] Oct 14 15:26:51 crc kubenswrapper[4945]: I1014 15:26:51.163912 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" event={"ID":"59c45798-dcb8-4fae-ba47-82a861d97023","Type":"ContainerStarted","Data":"049de480185805d49515ac5550a4fcd8abd5e154b5659c30a01697ee1a8090fa"} Oct 14 15:26:51 crc kubenswrapper[4945]: I1014 15:26:51.167422 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" event={"ID":"740d5a8e-2e4a-4fd1-8412-d5b989734f2f","Type":"ContainerStarted","Data":"413b3b6de4d72477369072be31273de6cf5d603d07f856b5ed88e073522d8d80"} Oct 14 15:26:55 crc kubenswrapper[4945]: I1014 15:26:55.206692 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" event={"ID":"740d5a8e-2e4a-4fd1-8412-d5b989734f2f","Type":"ContainerStarted","Data":"f1ab5d4d222241c4e1c7ac7b74a8b02a8d98f110887719fc451297884879f0a2"} Oct 14 15:26:55 crc kubenswrapper[4945]: I1014 15:26:55.207299 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:26:55 crc kubenswrapper[4945]: I1014 15:26:55.242304 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" podStartSLOduration=2.080820519 podStartE2EDuration="6.242288259s" podCreationTimestamp="2025-10-14 15:26:49 +0000 UTC" firstStartedPulling="2025-10-14 15:26:50.40706306 +0000 UTC m=+740.391111428" lastFinishedPulling="2025-10-14 15:26:54.5685308 +0000 UTC m=+744.552579168" observedRunningTime="2025-10-14 15:26:55.240159778 +0000 UTC m=+745.224208136" watchObservedRunningTime="2025-10-14 15:26:55.242288259 +0000 UTC m=+745.226336617" Oct 14 15:26:57 crc kubenswrapper[4945]: I1014 15:26:57.219084 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" event={"ID":"59c45798-dcb8-4fae-ba47-82a861d97023","Type":"ContainerStarted","Data":"bfe994746afb56e4c1d473eeba7e6a440c0521c11046dde9abd31bbac136e946"} Oct 14 15:26:57 crc kubenswrapper[4945]: I1014 15:26:57.219389 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:26:57 crc kubenswrapper[4945]: I1014 15:26:57.236021 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" podStartSLOduration=1.907794216 podStartE2EDuration="7.236006786s" podCreationTimestamp="2025-10-14 15:26:50 +0000 UTC" firstStartedPulling="2025-10-14 15:26:51.127608312 +0000 UTC m=+741.111656690" lastFinishedPulling="2025-10-14 15:26:56.455820892 +0000 UTC m=+746.439869260" observedRunningTime="2025-10-14 15:26:57.234462112 +0000 UTC m=+747.218510480" watchObservedRunningTime="2025-10-14 15:26:57.236006786 +0000 UTC m=+747.220055154" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.398080 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-glv7n"] Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.398750 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" podUID="58633adf-8733-4b72-a9da-78e278ebe5fd" containerName="controller-manager" containerID="cri-o://8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f" gracePeriod=30 Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.415672 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg"] Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.416068 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" podUID="7ff096c7-1a05-493a-9e19-8bec41303e6c" containerName="route-controller-manager" containerID="cri-o://1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae" gracePeriod=30 Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.853463 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.857122 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.865817 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-client-ca\") pod \"58633adf-8733-4b72-a9da-78e278ebe5fd\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.865861 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ff096c7-1a05-493a-9e19-8bec41303e6c-config\") pod \"7ff096c7-1a05-493a-9e19-8bec41303e6c\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.865960 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7ff096c7-1a05-493a-9e19-8bec41303e6c-client-ca\") pod \"7ff096c7-1a05-493a-9e19-8bec41303e6c\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.865990 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58633adf-8733-4b72-a9da-78e278ebe5fd-serving-cert\") pod \"58633adf-8733-4b72-a9da-78e278ebe5fd\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.866029 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff096c7-1a05-493a-9e19-8bec41303e6c-serving-cert\") pod \"7ff096c7-1a05-493a-9e19-8bec41303e6c\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.866072 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7jdq\" (UniqueName: \"kubernetes.io/projected/7ff096c7-1a05-493a-9e19-8bec41303e6c-kube-api-access-s7jdq\") pod \"7ff096c7-1a05-493a-9e19-8bec41303e6c\" (UID: \"7ff096c7-1a05-493a-9e19-8bec41303e6c\") " Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.866123 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-config\") pod \"58633adf-8733-4b72-a9da-78e278ebe5fd\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.866149 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-proxy-ca-bundles\") pod \"58633adf-8733-4b72-a9da-78e278ebe5fd\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.866214 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xq8ht\" (UniqueName: \"kubernetes.io/projected/58633adf-8733-4b72-a9da-78e278ebe5fd-kube-api-access-xq8ht\") pod \"58633adf-8733-4b72-a9da-78e278ebe5fd\" (UID: \"58633adf-8733-4b72-a9da-78e278ebe5fd\") " Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.866747 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-client-ca" (OuterVolumeSpecName: "client-ca") pod "58633adf-8733-4b72-a9da-78e278ebe5fd" (UID: "58633adf-8733-4b72-a9da-78e278ebe5fd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.866780 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff096c7-1a05-493a-9e19-8bec41303e6c-client-ca" (OuterVolumeSpecName: "client-ca") pod "7ff096c7-1a05-493a-9e19-8bec41303e6c" (UID: "7ff096c7-1a05-493a-9e19-8bec41303e6c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.866761 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff096c7-1a05-493a-9e19-8bec41303e6c-config" (OuterVolumeSpecName: "config") pod "7ff096c7-1a05-493a-9e19-8bec41303e6c" (UID: "7ff096c7-1a05-493a-9e19-8bec41303e6c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.867254 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-config" (OuterVolumeSpecName: "config") pod "58633adf-8733-4b72-a9da-78e278ebe5fd" (UID: "58633adf-8733-4b72-a9da-78e278ebe5fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.867856 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "58633adf-8733-4b72-a9da-78e278ebe5fd" (UID: "58633adf-8733-4b72-a9da-78e278ebe5fd"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.873132 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ff096c7-1a05-493a-9e19-8bec41303e6c-kube-api-access-s7jdq" (OuterVolumeSpecName: "kube-api-access-s7jdq") pod "7ff096c7-1a05-493a-9e19-8bec41303e6c" (UID: "7ff096c7-1a05-493a-9e19-8bec41303e6c"). InnerVolumeSpecName "kube-api-access-s7jdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.874155 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ff096c7-1a05-493a-9e19-8bec41303e6c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7ff096c7-1a05-493a-9e19-8bec41303e6c" (UID: "7ff096c7-1a05-493a-9e19-8bec41303e6c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.881150 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58633adf-8733-4b72-a9da-78e278ebe5fd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "58633adf-8733-4b72-a9da-78e278ebe5fd" (UID: "58633adf-8733-4b72-a9da-78e278ebe5fd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.882309 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58633adf-8733-4b72-a9da-78e278ebe5fd-kube-api-access-xq8ht" (OuterVolumeSpecName: "kube-api-access-xq8ht") pod "58633adf-8733-4b72-a9da-78e278ebe5fd" (UID: "58633adf-8733-4b72-a9da-78e278ebe5fd"). InnerVolumeSpecName "kube-api-access-xq8ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.968089 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ff096c7-1a05-493a-9e19-8bec41303e6c-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.968119 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7ff096c7-1a05-493a-9e19-8bec41303e6c-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.968130 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58633adf-8733-4b72-a9da-78e278ebe5fd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.968139 4945 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff096c7-1a05-493a-9e19-8bec41303e6c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.968148 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7jdq\" (UniqueName: \"kubernetes.io/projected/7ff096c7-1a05-493a-9e19-8bec41303e6c-kube-api-access-s7jdq\") on node \"crc\" DevicePath \"\"" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.968161 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.968169 4945 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.968177 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xq8ht\" (UniqueName: \"kubernetes.io/projected/58633adf-8733-4b72-a9da-78e278ebe5fd-kube-api-access-xq8ht\") on node \"crc\" DevicePath \"\"" Oct 14 15:27:06 crc kubenswrapper[4945]: I1014 15:27:06.968184 4945 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58633adf-8733-4b72-a9da-78e278ebe5fd-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.276144 4945 generic.go:334] "Generic (PLEG): container finished" podID="58633adf-8733-4b72-a9da-78e278ebe5fd" containerID="8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f" exitCode=0 Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.276192 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" event={"ID":"58633adf-8733-4b72-a9da-78e278ebe5fd","Type":"ContainerDied","Data":"8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f"} Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.276233 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.276263 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-glv7n" event={"ID":"58633adf-8733-4b72-a9da-78e278ebe5fd","Type":"ContainerDied","Data":"7a5966cb1f53b1aa26beed514ab86c478a3fac8bc1e07149cde7e14883b1b58f"} Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.276286 4945 scope.go:117] "RemoveContainer" containerID="8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f" Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.277689 4945 generic.go:334] "Generic (PLEG): container finished" podID="7ff096c7-1a05-493a-9e19-8bec41303e6c" containerID="1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae" exitCode=0 Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.277720 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.277739 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" event={"ID":"7ff096c7-1a05-493a-9e19-8bec41303e6c","Type":"ContainerDied","Data":"1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae"} Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.277772 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg" event={"ID":"7ff096c7-1a05-493a-9e19-8bec41303e6c","Type":"ContainerDied","Data":"205e4e3af51b98cfd46a1adf0e05e4332b649d95f8c3fe641d631defc95e1a37"} Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.305677 4945 scope.go:117] "RemoveContainer" containerID="8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f" Oct 14 15:27:07 crc kubenswrapper[4945]: E1014 15:27:07.308455 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f\": container with ID starting with 8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f not found: ID does not exist" containerID="8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f" Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.308494 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f"} err="failed to get container status \"8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f\": rpc error: code = NotFound desc = could not find container \"8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f\": container with ID starting with 8eb325cd0dbc092ebae011724fe1d4126ea35a397e1ec50878d49a896394691f not found: ID does not exist" Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.308515 4945 scope.go:117] "RemoveContainer" containerID="1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae" Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.311106 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-glv7n"] Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.314050 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-glv7n"] Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.320462 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg"] Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.325608 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hxtjg"] Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.333334 4945 scope.go:117] "RemoveContainer" containerID="1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae" Oct 14 15:27:07 crc kubenswrapper[4945]: E1014 15:27:07.333906 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae\": container with ID starting with 1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae not found: ID does not exist" containerID="1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae" Oct 14 15:27:07 crc kubenswrapper[4945]: I1014 15:27:07.333944 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae"} err="failed to get container status \"1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae\": rpc error: code = NotFound desc = could not find container \"1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae\": container with ID starting with 1bf2613f52a37a2d47531e7d17bc01ec52e4d8c22b407cb7e84b12d205c717ae not found: ID does not exist" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.321729 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb"] Oct 14 15:27:08 crc kubenswrapper[4945]: E1014 15:27:08.322124 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff096c7-1a05-493a-9e19-8bec41303e6c" containerName="route-controller-manager" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.322148 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff096c7-1a05-493a-9e19-8bec41303e6c" containerName="route-controller-manager" Oct 14 15:27:08 crc kubenswrapper[4945]: E1014 15:27:08.322173 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58633adf-8733-4b72-a9da-78e278ebe5fd" containerName="controller-manager" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.322186 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="58633adf-8733-4b72-a9da-78e278ebe5fd" containerName="controller-manager" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.322381 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="58633adf-8733-4b72-a9da-78e278ebe5fd" containerName="controller-manager" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.322430 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ff096c7-1a05-493a-9e19-8bec41303e6c" containerName="route-controller-manager" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.323107 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.326180 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b"] Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.326370 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.326459 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.326604 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.326771 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.326810 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.327025 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.327026 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.328802 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.329169 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.330103 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.330362 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.331030 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.332200 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.338761 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.343512 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b"] Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.347786 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb"] Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.384844 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-serving-cert\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.385010 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2vp6\" (UniqueName: \"kubernetes.io/projected/5d791869-2b1f-477b-a7a5-f779d0a147c2-kube-api-access-w2vp6\") pod \"route-controller-manager-bdb5bdc84-jzd6b\" (UID: \"5d791869-2b1f-477b-a7a5-f779d0a147c2\") " pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.385051 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-proxy-ca-bundles\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.385100 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d791869-2b1f-477b-a7a5-f779d0a147c2-client-ca\") pod \"route-controller-manager-bdb5bdc84-jzd6b\" (UID: \"5d791869-2b1f-477b-a7a5-f779d0a147c2\") " pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.385157 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-config\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.385197 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-client-ca\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.385286 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d791869-2b1f-477b-a7a5-f779d0a147c2-config\") pod \"route-controller-manager-bdb5bdc84-jzd6b\" (UID: \"5d791869-2b1f-477b-a7a5-f779d0a147c2\") " pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.385331 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dt6t\" (UniqueName: \"kubernetes.io/projected/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-kube-api-access-4dt6t\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.385459 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d791869-2b1f-477b-a7a5-f779d0a147c2-serving-cert\") pod \"route-controller-manager-bdb5bdc84-jzd6b\" (UID: \"5d791869-2b1f-477b-a7a5-f779d0a147c2\") " pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.486988 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-serving-cert\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.487084 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2vp6\" (UniqueName: \"kubernetes.io/projected/5d791869-2b1f-477b-a7a5-f779d0a147c2-kube-api-access-w2vp6\") pod \"route-controller-manager-bdb5bdc84-jzd6b\" (UID: \"5d791869-2b1f-477b-a7a5-f779d0a147c2\") " pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.487136 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-proxy-ca-bundles\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.487200 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d791869-2b1f-477b-a7a5-f779d0a147c2-client-ca\") pod \"route-controller-manager-bdb5bdc84-jzd6b\" (UID: \"5d791869-2b1f-477b-a7a5-f779d0a147c2\") " pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.487250 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-config\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.487308 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-client-ca\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.487379 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d791869-2b1f-477b-a7a5-f779d0a147c2-config\") pod \"route-controller-manager-bdb5bdc84-jzd6b\" (UID: \"5d791869-2b1f-477b-a7a5-f779d0a147c2\") " pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.487508 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dt6t\" (UniqueName: \"kubernetes.io/projected/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-kube-api-access-4dt6t\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.488153 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d791869-2b1f-477b-a7a5-f779d0a147c2-serving-cert\") pod \"route-controller-manager-bdb5bdc84-jzd6b\" (UID: \"5d791869-2b1f-477b-a7a5-f779d0a147c2\") " pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.488557 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d791869-2b1f-477b-a7a5-f779d0a147c2-client-ca\") pod \"route-controller-manager-bdb5bdc84-jzd6b\" (UID: \"5d791869-2b1f-477b-a7a5-f779d0a147c2\") " pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.488701 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-client-ca\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.489048 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-config\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.489174 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d791869-2b1f-477b-a7a5-f779d0a147c2-config\") pod \"route-controller-manager-bdb5bdc84-jzd6b\" (UID: \"5d791869-2b1f-477b-a7a5-f779d0a147c2\") " pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.490170 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-proxy-ca-bundles\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.492581 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d791869-2b1f-477b-a7a5-f779d0a147c2-serving-cert\") pod \"route-controller-manager-bdb5bdc84-jzd6b\" (UID: \"5d791869-2b1f-477b-a7a5-f779d0a147c2\") " pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.505733 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-serving-cert\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.516729 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2vp6\" (UniqueName: \"kubernetes.io/projected/5d791869-2b1f-477b-a7a5-f779d0a147c2-kube-api-access-w2vp6\") pod \"route-controller-manager-bdb5bdc84-jzd6b\" (UID: \"5d791869-2b1f-477b-a7a5-f779d0a147c2\") " pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.517018 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dt6t\" (UniqueName: \"kubernetes.io/projected/e9ff1709-d742-4ad1-9c33-9546b0a6b0b7-kube-api-access-4dt6t\") pod \"controller-manager-5547c6cbd5-m7cbb\" (UID: \"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7\") " pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.645037 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.652056 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.776537 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58633adf-8733-4b72-a9da-78e278ebe5fd" path="/var/lib/kubelet/pods/58633adf-8733-4b72-a9da-78e278ebe5fd/volumes" Oct 14 15:27:08 crc kubenswrapper[4945]: I1014 15:27:08.777282 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ff096c7-1a05-493a-9e19-8bec41303e6c" path="/var/lib/kubelet/pods/7ff096c7-1a05-493a-9e19-8bec41303e6c/volumes" Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.046769 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb"] Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.110112 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b"] Oct 14 15:27:09 crc kubenswrapper[4945]: W1014 15:27:09.132620 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d791869_2b1f_477b_a7a5_f779d0a147c2.slice/crio-f77a78bd8af2018ea0fa712e2a21081d6e38394d350ff38ad1ec58904d9a15e3 WatchSource:0}: Error finding container f77a78bd8af2018ea0fa712e2a21081d6e38394d350ff38ad1ec58904d9a15e3: Status 404 returned error can't find the container with id f77a78bd8af2018ea0fa712e2a21081d6e38394d350ff38ad1ec58904d9a15e3 Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.291882 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" event={"ID":"5d791869-2b1f-477b-a7a5-f779d0a147c2","Type":"ContainerStarted","Data":"028513af5a2bbafd9f3ca019d4e0577f6fef888f4504de195288b12adc8a827e"} Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.291916 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" event={"ID":"5d791869-2b1f-477b-a7a5-f779d0a147c2","Type":"ContainerStarted","Data":"f77a78bd8af2018ea0fa712e2a21081d6e38394d350ff38ad1ec58904d9a15e3"} Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.292778 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.293836 4945 patch_prober.go:28] interesting pod/route-controller-manager-bdb5bdc84-jzd6b container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.50:8443/healthz\": dial tcp 10.217.0.50:8443: connect: connection refused" start-of-body= Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.293905 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" podUID="5d791869-2b1f-477b-a7a5-f779d0a147c2" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.50:8443/healthz\": dial tcp 10.217.0.50:8443: connect: connection refused" Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.296415 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" event={"ID":"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7","Type":"ContainerStarted","Data":"bb0524162a13fc2634b2aac53774618b154a65a55baf2be5acb836bd877cc77c"} Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.296461 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" event={"ID":"e9ff1709-d742-4ad1-9c33-9546b0a6b0b7","Type":"ContainerStarted","Data":"8c37d854ad6ae1561447154dcf0ba57678df63eb7d250f97f985bcb725e49c4c"} Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.296974 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.297728 4945 patch_prober.go:28] interesting pod/controller-manager-5547c6cbd5-m7cbb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: connect: connection refused" start-of-body= Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.297781 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" podUID="e9ff1709-d742-4ad1-9c33-9546b0a6b0b7" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: connect: connection refused" Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.312169 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" podStartSLOduration=3.312152077 podStartE2EDuration="3.312152077s" podCreationTimestamp="2025-10-14 15:27:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:27:09.310816579 +0000 UTC m=+759.294864957" watchObservedRunningTime="2025-10-14 15:27:09.312152077 +0000 UTC m=+759.296200455" Oct 14 15:27:09 crc kubenswrapper[4945]: I1014 15:27:09.332668 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" podStartSLOduration=3.332653135 podStartE2EDuration="3.332653135s" podCreationTimestamp="2025-10-14 15:27:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:27:09.330627147 +0000 UTC m=+759.314675505" watchObservedRunningTime="2025-10-14 15:27:09.332653135 +0000 UTC m=+759.316701503" Oct 14 15:27:10 crc kubenswrapper[4945]: I1014 15:27:10.305500 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5547c6cbd5-m7cbb" Oct 14 15:27:10 crc kubenswrapper[4945]: I1014 15:27:10.305838 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-bdb5bdc84-jzd6b" Oct 14 15:27:10 crc kubenswrapper[4945]: I1014 15:27:10.577172 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6545bcf4ff-f248w" Oct 14 15:27:15 crc kubenswrapper[4945]: I1014 15:27:15.971597 4945 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.106973 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-9f47f9675-47zvd" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.870942 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-h88zg"] Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.876929 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.879569 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-2t6tf" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.881483 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.882097 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.886118 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9"] Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.886943 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.891701 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9"] Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.901573 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.908025 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-metrics\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.908081 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-frr-conf\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.908099 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-metrics-certs\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.908120 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6767\" (UniqueName: \"kubernetes.io/projected/79b86a9d-a49c-428a-b3f9-3e2625ad1abc-kube-api-access-f6767\") pod \"frr-k8s-webhook-server-64bf5d555-pfbt9\" (UID: \"79b86a9d-a49c-428a-b3f9-3e2625ad1abc\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.908187 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-reloader\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.908225 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-frr-startup\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.908252 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b86a9d-a49c-428a-b3f9-3e2625ad1abc-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pfbt9\" (UID: \"79b86a9d-a49c-428a-b3f9-3e2625ad1abc\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.908269 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpxbv\" (UniqueName: \"kubernetes.io/projected/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-kube-api-access-cpxbv\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.908309 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-frr-sockets\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.979775 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-lgvg5"] Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.980678 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-lgvg5" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.983035 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-tvn99"] Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.983770 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.984747 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.985031 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.985228 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.985298 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-bdqjw" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.989725 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 14 15:27:30 crc kubenswrapper[4945]: I1014 15:27:30.996722 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-tvn99"] Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.008962 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnrr2\" (UniqueName: \"kubernetes.io/projected/9a07810a-86fd-4dd6-a704-037bea9afcab-kube-api-access-vnrr2\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.009101 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/9a07810a-86fd-4dd6-a704-037bea9afcab-metallb-excludel2\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.009218 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-reloader\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.009340 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-frr-startup\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.009435 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a07810a-86fd-4dd6-a704-037bea9afcab-metrics-certs\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.009541 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38c1e540-b577-472f-82e3-5079b3823d87-metrics-certs\") pod \"controller-68d546b9d8-tvn99\" (UID: \"38c1e540-b577-472f-82e3-5079b3823d87\") " pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.009630 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b86a9d-a49c-428a-b3f9-3e2625ad1abc-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pfbt9\" (UID: \"79b86a9d-a49c-428a-b3f9-3e2625ad1abc\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.009720 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpxbv\" (UniqueName: \"kubernetes.io/projected/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-kube-api-access-cpxbv\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: E1014 15:27:31.009825 4945 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 14 15:27:31 crc kubenswrapper[4945]: E1014 15:27:31.009917 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79b86a9d-a49c-428a-b3f9-3e2625ad1abc-cert podName:79b86a9d-a49c-428a-b3f9-3e2625ad1abc nodeName:}" failed. No retries permitted until 2025-10-14 15:27:31.509899109 +0000 UTC m=+781.493947477 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79b86a9d-a49c-428a-b3f9-3e2625ad1abc-cert") pod "frr-k8s-webhook-server-64bf5d555-pfbt9" (UID: "79b86a9d-a49c-428a-b3f9-3e2625ad1abc") : secret "frr-k8s-webhook-server-cert" not found Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.009653 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-reloader\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.009829 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9a07810a-86fd-4dd6-a704-037bea9afcab-memberlist\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.010005 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/38c1e540-b577-472f-82e3-5079b3823d87-cert\") pod \"controller-68d546b9d8-tvn99\" (UID: \"38c1e540-b577-472f-82e3-5079b3823d87\") " pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.010066 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-frr-sockets\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.010089 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-metrics\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.010129 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c2bv\" (UniqueName: \"kubernetes.io/projected/38c1e540-b577-472f-82e3-5079b3823d87-kube-api-access-4c2bv\") pod \"controller-68d546b9d8-tvn99\" (UID: \"38c1e540-b577-472f-82e3-5079b3823d87\") " pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.010156 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-frr-conf\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.010173 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-metrics-certs\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.010193 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6767\" (UniqueName: \"kubernetes.io/projected/79b86a9d-a49c-428a-b3f9-3e2625ad1abc-kube-api-access-f6767\") pod \"frr-k8s-webhook-server-64bf5d555-pfbt9\" (UID: \"79b86a9d-a49c-428a-b3f9-3e2625ad1abc\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.010357 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-frr-startup\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.010564 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-metrics\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: E1014 15:27:31.010649 4945 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 14 15:27:31 crc kubenswrapper[4945]: E1014 15:27:31.010697 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-metrics-certs podName:2b3404d8-44ad-4e9c-9c35-1d65d401b9cb nodeName:}" failed. No retries permitted until 2025-10-14 15:27:31.510682232 +0000 UTC m=+781.494730600 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-metrics-certs") pod "frr-k8s-h88zg" (UID: "2b3404d8-44ad-4e9c-9c35-1d65d401b9cb") : secret "frr-k8s-certs-secret" not found Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.010701 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-frr-conf\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.010993 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-frr-sockets\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.028345 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpxbv\" (UniqueName: \"kubernetes.io/projected/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-kube-api-access-cpxbv\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.029618 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6767\" (UniqueName: \"kubernetes.io/projected/79b86a9d-a49c-428a-b3f9-3e2625ad1abc-kube-api-access-f6767\") pod \"frr-k8s-webhook-server-64bf5d555-pfbt9\" (UID: \"79b86a9d-a49c-428a-b3f9-3e2625ad1abc\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.111410 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/9a07810a-86fd-4dd6-a704-037bea9afcab-metallb-excludel2\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.111464 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a07810a-86fd-4dd6-a704-037bea9afcab-metrics-certs\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.111490 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38c1e540-b577-472f-82e3-5079b3823d87-metrics-certs\") pod \"controller-68d546b9d8-tvn99\" (UID: \"38c1e540-b577-472f-82e3-5079b3823d87\") " pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.111531 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9a07810a-86fd-4dd6-a704-037bea9afcab-memberlist\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.111555 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/38c1e540-b577-472f-82e3-5079b3823d87-cert\") pod \"controller-68d546b9d8-tvn99\" (UID: \"38c1e540-b577-472f-82e3-5079b3823d87\") " pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.111588 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c2bv\" (UniqueName: \"kubernetes.io/projected/38c1e540-b577-472f-82e3-5079b3823d87-kube-api-access-4c2bv\") pod \"controller-68d546b9d8-tvn99\" (UID: \"38c1e540-b577-472f-82e3-5079b3823d87\") " pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.111615 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnrr2\" (UniqueName: \"kubernetes.io/projected/9a07810a-86fd-4dd6-a704-037bea9afcab-kube-api-access-vnrr2\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.112779 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/9a07810a-86fd-4dd6-a704-037bea9afcab-metallb-excludel2\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:31 crc kubenswrapper[4945]: E1014 15:27:31.112850 4945 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 14 15:27:31 crc kubenswrapper[4945]: E1014 15:27:31.112926 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a07810a-86fd-4dd6-a704-037bea9afcab-memberlist podName:9a07810a-86fd-4dd6-a704-037bea9afcab nodeName:}" failed. No retries permitted until 2025-10-14 15:27:31.61291424 +0000 UTC m=+781.596962608 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/9a07810a-86fd-4dd6-a704-037bea9afcab-memberlist") pod "speaker-lgvg5" (UID: "9a07810a-86fd-4dd6-a704-037bea9afcab") : secret "metallb-memberlist" not found Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.116474 4945 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.118658 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38c1e540-b577-472f-82e3-5079b3823d87-metrics-certs\") pod \"controller-68d546b9d8-tvn99\" (UID: \"38c1e540-b577-472f-82e3-5079b3823d87\") " pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.118665 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a07810a-86fd-4dd6-a704-037bea9afcab-metrics-certs\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.127598 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/38c1e540-b577-472f-82e3-5079b3823d87-cert\") pod \"controller-68d546b9d8-tvn99\" (UID: \"38c1e540-b577-472f-82e3-5079b3823d87\") " pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.130808 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c2bv\" (UniqueName: \"kubernetes.io/projected/38c1e540-b577-472f-82e3-5079b3823d87-kube-api-access-4c2bv\") pod \"controller-68d546b9d8-tvn99\" (UID: \"38c1e540-b577-472f-82e3-5079b3823d87\") " pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.133633 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnrr2\" (UniqueName: \"kubernetes.io/projected/9a07810a-86fd-4dd6-a704-037bea9afcab-kube-api-access-vnrr2\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.324747 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.515624 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-metrics-certs\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.515714 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b86a9d-a49c-428a-b3f9-3e2625ad1abc-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pfbt9\" (UID: \"79b86a9d-a49c-428a-b3f9-3e2625ad1abc\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.519143 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b3404d8-44ad-4e9c-9c35-1d65d401b9cb-metrics-certs\") pod \"frr-k8s-h88zg\" (UID: \"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb\") " pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.519172 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b86a9d-a49c-428a-b3f9-3e2625ad1abc-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pfbt9\" (UID: \"79b86a9d-a49c-428a-b3f9-3e2625ad1abc\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.524138 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.625640 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9a07810a-86fd-4dd6-a704-037bea9afcab-memberlist\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:31 crc kubenswrapper[4945]: E1014 15:27:31.625779 4945 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 14 15:27:31 crc kubenswrapper[4945]: E1014 15:27:31.625829 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a07810a-86fd-4dd6-a704-037bea9afcab-memberlist podName:9a07810a-86fd-4dd6-a704-037bea9afcab nodeName:}" failed. No retries permitted until 2025-10-14 15:27:32.625816292 +0000 UTC m=+782.609864660 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/9a07810a-86fd-4dd6-a704-037bea9afcab-memberlist") pod "speaker-lgvg5" (UID: "9a07810a-86fd-4dd6-a704-037bea9afcab") : secret "metallb-memberlist" not found Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.742377 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-tvn99"] Oct 14 15:27:31 crc kubenswrapper[4945]: W1014 15:27:31.750175 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38c1e540_b577_472f_82e3_5079b3823d87.slice/crio-596cef094ee2e9b209cabb83f5681b89a77586caebd04615a0f6675c217e0695 WatchSource:0}: Error finding container 596cef094ee2e9b209cabb83f5681b89a77586caebd04615a0f6675c217e0695: Status 404 returned error can't find the container with id 596cef094ee2e9b209cabb83f5681b89a77586caebd04615a0f6675c217e0695 Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.806973 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:31 crc kubenswrapper[4945]: I1014 15:27:31.966441 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9"] Oct 14 15:27:31 crc kubenswrapper[4945]: W1014 15:27:31.973266 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79b86a9d_a49c_428a_b3f9_3e2625ad1abc.slice/crio-a0c5b8e11a28e9a86f1d44bb4d98a113da2525fb58ab7e307a99897e34ee5b38 WatchSource:0}: Error finding container a0c5b8e11a28e9a86f1d44bb4d98a113da2525fb58ab7e307a99897e34ee5b38: Status 404 returned error can't find the container with id a0c5b8e11a28e9a86f1d44bb4d98a113da2525fb58ab7e307a99897e34ee5b38 Oct 14 15:27:32 crc kubenswrapper[4945]: I1014 15:27:32.438963 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" event={"ID":"79b86a9d-a49c-428a-b3f9-3e2625ad1abc","Type":"ContainerStarted","Data":"a0c5b8e11a28e9a86f1d44bb4d98a113da2525fb58ab7e307a99897e34ee5b38"} Oct 14 15:27:32 crc kubenswrapper[4945]: I1014 15:27:32.440436 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h88zg" event={"ID":"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb","Type":"ContainerStarted","Data":"640d24570644a5ffaf3e2c9a1caccc636e277851b3f08db6aade397ac4d248b1"} Oct 14 15:27:32 crc kubenswrapper[4945]: I1014 15:27:32.442172 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-tvn99" event={"ID":"38c1e540-b577-472f-82e3-5079b3823d87","Type":"ContainerStarted","Data":"46ac2f0fe13cb670d56da10663a47c0e9f3b5e08e57f4fa29dc0a26913f4fb0f"} Oct 14 15:27:32 crc kubenswrapper[4945]: I1014 15:27:32.442205 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-tvn99" event={"ID":"38c1e540-b577-472f-82e3-5079b3823d87","Type":"ContainerStarted","Data":"22399296bf76c87a0fcd7bfdba0f95b2db4703bb17e5ea1d03fd32583b1ad864"} Oct 14 15:27:32 crc kubenswrapper[4945]: I1014 15:27:32.442223 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-tvn99" event={"ID":"38c1e540-b577-472f-82e3-5079b3823d87","Type":"ContainerStarted","Data":"596cef094ee2e9b209cabb83f5681b89a77586caebd04615a0f6675c217e0695"} Oct 14 15:27:32 crc kubenswrapper[4945]: I1014 15:27:32.442475 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:32 crc kubenswrapper[4945]: I1014 15:27:32.465905 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-tvn99" podStartSLOduration=2.465863291 podStartE2EDuration="2.465863291s" podCreationTimestamp="2025-10-14 15:27:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:27:32.460648548 +0000 UTC m=+782.444696936" watchObservedRunningTime="2025-10-14 15:27:32.465863291 +0000 UTC m=+782.449911659" Oct 14 15:27:32 crc kubenswrapper[4945]: I1014 15:27:32.651897 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9a07810a-86fd-4dd6-a704-037bea9afcab-memberlist\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:32 crc kubenswrapper[4945]: I1014 15:27:32.660523 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9a07810a-86fd-4dd6-a704-037bea9afcab-memberlist\") pod \"speaker-lgvg5\" (UID: \"9a07810a-86fd-4dd6-a704-037bea9afcab\") " pod="metallb-system/speaker-lgvg5" Oct 14 15:27:32 crc kubenswrapper[4945]: I1014 15:27:32.796003 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-lgvg5" Oct 14 15:27:33 crc kubenswrapper[4945]: I1014 15:27:33.471906 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-lgvg5" event={"ID":"9a07810a-86fd-4dd6-a704-037bea9afcab","Type":"ContainerStarted","Data":"d61731be7106f6b3971ecd812141c93fc137cf9a43947546fa9b4674a785a0e2"} Oct 14 15:27:33 crc kubenswrapper[4945]: I1014 15:27:33.472274 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-lgvg5" event={"ID":"9a07810a-86fd-4dd6-a704-037bea9afcab","Type":"ContainerStarted","Data":"817b8ec62d67aaf71816124568c5b6e6d6a7c0ee0a970870ffecb49e46acd799"} Oct 14 15:27:33 crc kubenswrapper[4945]: I1014 15:27:33.472295 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-lgvg5" event={"ID":"9a07810a-86fd-4dd6-a704-037bea9afcab","Type":"ContainerStarted","Data":"37a688d6a244f5120e36261943b8270615d9f0e670f0193279286558c053b689"} Oct 14 15:27:33 crc kubenswrapper[4945]: I1014 15:27:33.472439 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-lgvg5" Oct 14 15:27:33 crc kubenswrapper[4945]: I1014 15:27:33.496851 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-lgvg5" podStartSLOduration=3.4968268780000002 podStartE2EDuration="3.496826878s" podCreationTimestamp="2025-10-14 15:27:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:27:33.4918245 +0000 UTC m=+783.475872868" watchObservedRunningTime="2025-10-14 15:27:33.496826878 +0000 UTC m=+783.480875246" Oct 14 15:27:39 crc kubenswrapper[4945]: I1014 15:27:39.516938 4945 generic.go:334] "Generic (PLEG): container finished" podID="2b3404d8-44ad-4e9c-9c35-1d65d401b9cb" containerID="861eef8b20ccaf34b11c0e49b263279764a5f101e64e4fe4f8f1a57909e8aaf9" exitCode=0 Oct 14 15:27:39 crc kubenswrapper[4945]: I1014 15:27:39.517013 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h88zg" event={"ID":"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb","Type":"ContainerDied","Data":"861eef8b20ccaf34b11c0e49b263279764a5f101e64e4fe4f8f1a57909e8aaf9"} Oct 14 15:27:39 crc kubenswrapper[4945]: I1014 15:27:39.520161 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" event={"ID":"79b86a9d-a49c-428a-b3f9-3e2625ad1abc","Type":"ContainerStarted","Data":"4289c4ecb90422e0f1d84a6e1cfe7b8fdd4bed93b22f144a04d512bc6d4c7247"} Oct 14 15:27:39 crc kubenswrapper[4945]: I1014 15:27:39.520427 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" Oct 14 15:27:39 crc kubenswrapper[4945]: I1014 15:27:39.554611 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" podStartSLOduration=2.540038642 podStartE2EDuration="9.554588454s" podCreationTimestamp="2025-10-14 15:27:30 +0000 UTC" firstStartedPulling="2025-10-14 15:27:31.975217118 +0000 UTC m=+781.959265496" lastFinishedPulling="2025-10-14 15:27:38.98976694 +0000 UTC m=+788.973815308" observedRunningTime="2025-10-14 15:27:39.553649626 +0000 UTC m=+789.537698024" watchObservedRunningTime="2025-10-14 15:27:39.554588454 +0000 UTC m=+789.538636822" Oct 14 15:27:40 crc kubenswrapper[4945]: I1014 15:27:40.528586 4945 generic.go:334] "Generic (PLEG): container finished" podID="2b3404d8-44ad-4e9c-9c35-1d65d401b9cb" containerID="edc0ce7fd1043345c56c3db2b0f3d6f3dcd920c0ac6ac43325b692954487a4d2" exitCode=0 Oct 14 15:27:40 crc kubenswrapper[4945]: I1014 15:27:40.528667 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h88zg" event={"ID":"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb","Type":"ContainerDied","Data":"edc0ce7fd1043345c56c3db2b0f3d6f3dcd920c0ac6ac43325b692954487a4d2"} Oct 14 15:27:41 crc kubenswrapper[4945]: I1014 15:27:41.538652 4945 generic.go:334] "Generic (PLEG): container finished" podID="2b3404d8-44ad-4e9c-9c35-1d65d401b9cb" containerID="cbf4ebeca0417ea761022fa48c6a0586e4498f05d4776df40f6fbe664f270bce" exitCode=0 Oct 14 15:27:41 crc kubenswrapper[4945]: I1014 15:27:41.538692 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h88zg" event={"ID":"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb","Type":"ContainerDied","Data":"cbf4ebeca0417ea761022fa48c6a0586e4498f05d4776df40f6fbe664f270bce"} Oct 14 15:27:42 crc kubenswrapper[4945]: I1014 15:27:42.547660 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h88zg" event={"ID":"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb","Type":"ContainerStarted","Data":"93053eb2772d324ea978a68870f18cc3922b058dbc4cc792224e4b7d56bee0cb"} Oct 14 15:27:42 crc kubenswrapper[4945]: I1014 15:27:42.547701 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h88zg" event={"ID":"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb","Type":"ContainerStarted","Data":"0fb8cba322b9d5f8612f266bb5b018daa58f45186741145c31473fbf666f13d0"} Oct 14 15:27:42 crc kubenswrapper[4945]: I1014 15:27:42.547711 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h88zg" event={"ID":"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb","Type":"ContainerStarted","Data":"e2e4f7ea579c42ebd45e1e8466548fc5ee4018335a0ab587a19b1dd8cc61ae4f"} Oct 14 15:27:43 crc kubenswrapper[4945]: I1014 15:27:43.555532 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h88zg" event={"ID":"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb","Type":"ContainerStarted","Data":"d1acb3739072a6369d87b389c6ccf07f1402d8319fc030d06bcb8342f6158cd9"} Oct 14 15:27:43 crc kubenswrapper[4945]: I1014 15:27:43.555761 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h88zg" event={"ID":"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb","Type":"ContainerStarted","Data":"6e8b79dbd11625c18a4cfc5f71cd90d6217212f3b62b895911cf0a7132c5766f"} Oct 14 15:27:43 crc kubenswrapper[4945]: I1014 15:27:43.555867 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:43 crc kubenswrapper[4945]: I1014 15:27:43.555948 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h88zg" event={"ID":"2b3404d8-44ad-4e9c-9c35-1d65d401b9cb","Type":"ContainerStarted","Data":"73d8461a88c1f66ac0f4eedcaf4a73b50d8cf0a4c2bd2da6a887498f372d00d1"} Oct 14 15:27:43 crc kubenswrapper[4945]: I1014 15:27:43.582987 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-h88zg" podStartSLOduration=6.702909506 podStartE2EDuration="13.582962939s" podCreationTimestamp="2025-10-14 15:27:30 +0000 UTC" firstStartedPulling="2025-10-14 15:27:32.119566768 +0000 UTC m=+782.103615126" lastFinishedPulling="2025-10-14 15:27:38.999620191 +0000 UTC m=+788.983668559" observedRunningTime="2025-10-14 15:27:43.579823056 +0000 UTC m=+793.563871424" watchObservedRunningTime="2025-10-14 15:27:43.582962939 +0000 UTC m=+793.567011307" Oct 14 15:27:46 crc kubenswrapper[4945]: I1014 15:27:46.795611 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:27:46 crc kubenswrapper[4945]: I1014 15:27:46.796016 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:27:46 crc kubenswrapper[4945]: I1014 15:27:46.807806 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:46 crc kubenswrapper[4945]: I1014 15:27:46.843555 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.220247 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4c9kk"] Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.223419 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.230841 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4c9kk"] Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.317152 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac0ee420-6ad5-41cd-aad0-64777519ecda-utilities\") pod \"community-operators-4c9kk\" (UID: \"ac0ee420-6ad5-41cd-aad0-64777519ecda\") " pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.317482 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac0ee420-6ad5-41cd-aad0-64777519ecda-catalog-content\") pod \"community-operators-4c9kk\" (UID: \"ac0ee420-6ad5-41cd-aad0-64777519ecda\") " pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.317507 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9cs8\" (UniqueName: \"kubernetes.io/projected/ac0ee420-6ad5-41cd-aad0-64777519ecda-kube-api-access-t9cs8\") pod \"community-operators-4c9kk\" (UID: \"ac0ee420-6ad5-41cd-aad0-64777519ecda\") " pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.331787 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-tvn99" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.419103 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac0ee420-6ad5-41cd-aad0-64777519ecda-utilities\") pod \"community-operators-4c9kk\" (UID: \"ac0ee420-6ad5-41cd-aad0-64777519ecda\") " pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.419209 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac0ee420-6ad5-41cd-aad0-64777519ecda-catalog-content\") pod \"community-operators-4c9kk\" (UID: \"ac0ee420-6ad5-41cd-aad0-64777519ecda\") " pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.419229 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9cs8\" (UniqueName: \"kubernetes.io/projected/ac0ee420-6ad5-41cd-aad0-64777519ecda-kube-api-access-t9cs8\") pod \"community-operators-4c9kk\" (UID: \"ac0ee420-6ad5-41cd-aad0-64777519ecda\") " pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.419632 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac0ee420-6ad5-41cd-aad0-64777519ecda-utilities\") pod \"community-operators-4c9kk\" (UID: \"ac0ee420-6ad5-41cd-aad0-64777519ecda\") " pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.419766 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac0ee420-6ad5-41cd-aad0-64777519ecda-catalog-content\") pod \"community-operators-4c9kk\" (UID: \"ac0ee420-6ad5-41cd-aad0-64777519ecda\") " pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.443725 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9cs8\" (UniqueName: \"kubernetes.io/projected/ac0ee420-6ad5-41cd-aad0-64777519ecda-kube-api-access-t9cs8\") pod \"community-operators-4c9kk\" (UID: \"ac0ee420-6ad5-41cd-aad0-64777519ecda\") " pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.536324 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pfbt9" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.555628 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:27:51 crc kubenswrapper[4945]: I1014 15:27:51.824979 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-h88zg" Oct 14 15:27:52 crc kubenswrapper[4945]: I1014 15:27:52.039055 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4c9kk"] Oct 14 15:27:52 crc kubenswrapper[4945]: I1014 15:27:52.626269 4945 generic.go:334] "Generic (PLEG): container finished" podID="ac0ee420-6ad5-41cd-aad0-64777519ecda" containerID="cd144878953a03d1d49d22bdc42bc6546f92e47729884983d06b0fc8f1f9ae45" exitCode=0 Oct 14 15:27:52 crc kubenswrapper[4945]: I1014 15:27:52.626344 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4c9kk" event={"ID":"ac0ee420-6ad5-41cd-aad0-64777519ecda","Type":"ContainerDied","Data":"cd144878953a03d1d49d22bdc42bc6546f92e47729884983d06b0fc8f1f9ae45"} Oct 14 15:27:52 crc kubenswrapper[4945]: I1014 15:27:52.626387 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4c9kk" event={"ID":"ac0ee420-6ad5-41cd-aad0-64777519ecda","Type":"ContainerStarted","Data":"8c645b2a26c4623db67f81132c8286733aaccdddbdcb60387b2f3e98df2e4ea5"} Oct 14 15:27:52 crc kubenswrapper[4945]: I1014 15:27:52.804391 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-lgvg5" Oct 14 15:27:54 crc kubenswrapper[4945]: I1014 15:27:54.641409 4945 generic.go:334] "Generic (PLEG): container finished" podID="ac0ee420-6ad5-41cd-aad0-64777519ecda" containerID="7c7085ab41f62433d1cecdc71acdd037f65c89b8bb46b5924169fd3bc8705adb" exitCode=0 Oct 14 15:27:54 crc kubenswrapper[4945]: I1014 15:27:54.641478 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4c9kk" event={"ID":"ac0ee420-6ad5-41cd-aad0-64777519ecda","Type":"ContainerDied","Data":"7c7085ab41f62433d1cecdc71acdd037f65c89b8bb46b5924169fd3bc8705adb"} Oct 14 15:27:55 crc kubenswrapper[4945]: I1014 15:27:55.650132 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4c9kk" event={"ID":"ac0ee420-6ad5-41cd-aad0-64777519ecda","Type":"ContainerStarted","Data":"5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd"} Oct 14 15:27:55 crc kubenswrapper[4945]: I1014 15:27:55.667541 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4c9kk" podStartSLOduration=1.825424134 podStartE2EDuration="4.667524028s" podCreationTimestamp="2025-10-14 15:27:51 +0000 UTC" firstStartedPulling="2025-10-14 15:27:52.629252103 +0000 UTC m=+802.613300521" lastFinishedPulling="2025-10-14 15:27:55.471352047 +0000 UTC m=+805.455400415" observedRunningTime="2025-10-14 15:27:55.665177609 +0000 UTC m=+805.649225977" watchObservedRunningTime="2025-10-14 15:27:55.667524028 +0000 UTC m=+805.651572396" Oct 14 15:27:58 crc kubenswrapper[4945]: I1014 15:27:58.191281 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-f9h5q"] Oct 14 15:27:58 crc kubenswrapper[4945]: I1014 15:27:58.192266 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-f9h5q" Oct 14 15:27:58 crc kubenswrapper[4945]: I1014 15:27:58.194952 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 14 15:27:58 crc kubenswrapper[4945]: I1014 15:27:58.195025 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 14 15:27:58 crc kubenswrapper[4945]: I1014 15:27:58.195255 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-vthvk" Oct 14 15:27:58 crc kubenswrapper[4945]: I1014 15:27:58.199199 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-f9h5q"] Oct 14 15:27:58 crc kubenswrapper[4945]: I1014 15:27:58.302340 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27x2r\" (UniqueName: \"kubernetes.io/projected/4b5016e5-1067-4b63-9db4-f5257354538d-kube-api-access-27x2r\") pod \"openstack-operator-index-f9h5q\" (UID: \"4b5016e5-1067-4b63-9db4-f5257354538d\") " pod="openstack-operators/openstack-operator-index-f9h5q" Oct 14 15:27:58 crc kubenswrapper[4945]: I1014 15:27:58.403465 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27x2r\" (UniqueName: \"kubernetes.io/projected/4b5016e5-1067-4b63-9db4-f5257354538d-kube-api-access-27x2r\") pod \"openstack-operator-index-f9h5q\" (UID: \"4b5016e5-1067-4b63-9db4-f5257354538d\") " pod="openstack-operators/openstack-operator-index-f9h5q" Oct 14 15:27:58 crc kubenswrapper[4945]: I1014 15:27:58.421764 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27x2r\" (UniqueName: \"kubernetes.io/projected/4b5016e5-1067-4b63-9db4-f5257354538d-kube-api-access-27x2r\") pod \"openstack-operator-index-f9h5q\" (UID: \"4b5016e5-1067-4b63-9db4-f5257354538d\") " pod="openstack-operators/openstack-operator-index-f9h5q" Oct 14 15:27:58 crc kubenswrapper[4945]: I1014 15:27:58.570997 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-f9h5q" Oct 14 15:27:58 crc kubenswrapper[4945]: I1014 15:27:58.962262 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-f9h5q"] Oct 14 15:27:58 crc kubenswrapper[4945]: W1014 15:27:58.972765 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b5016e5_1067_4b63_9db4_f5257354538d.slice/crio-466c286fda3231e641af82e30128c065f3e0a65037aee65273dbfefbf1897c58 WatchSource:0}: Error finding container 466c286fda3231e641af82e30128c065f3e0a65037aee65273dbfefbf1897c58: Status 404 returned error can't find the container with id 466c286fda3231e641af82e30128c065f3e0a65037aee65273dbfefbf1897c58 Oct 14 15:27:59 crc kubenswrapper[4945]: I1014 15:27:59.674030 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-f9h5q" event={"ID":"4b5016e5-1067-4b63-9db4-f5257354538d","Type":"ContainerStarted","Data":"466c286fda3231e641af82e30128c065f3e0a65037aee65273dbfefbf1897c58"} Oct 14 15:28:01 crc kubenswrapper[4945]: I1014 15:28:01.557087 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:28:01 crc kubenswrapper[4945]: I1014 15:28:01.557574 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:28:01 crc kubenswrapper[4945]: I1014 15:28:01.620026 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:28:01 crc kubenswrapper[4945]: I1014 15:28:01.685726 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-f9h5q" event={"ID":"4b5016e5-1067-4b63-9db4-f5257354538d","Type":"ContainerStarted","Data":"5eb295af737237af672f902de17535fe49136c8bc93f5a281ddf69d0922cd8e4"} Oct 14 15:28:01 crc kubenswrapper[4945]: I1014 15:28:01.710654 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-f9h5q" podStartSLOduration=1.684773765 podStartE2EDuration="3.710632102s" podCreationTimestamp="2025-10-14 15:27:58 +0000 UTC" firstStartedPulling="2025-10-14 15:27:58.975584498 +0000 UTC m=+808.959632866" lastFinishedPulling="2025-10-14 15:28:01.001442825 +0000 UTC m=+810.985491203" observedRunningTime="2025-10-14 15:28:01.709219241 +0000 UTC m=+811.693267639" watchObservedRunningTime="2025-10-14 15:28:01.710632102 +0000 UTC m=+811.694680490" Oct 14 15:28:01 crc kubenswrapper[4945]: I1014 15:28:01.728453 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:28:05 crc kubenswrapper[4945]: I1014 15:28:05.389786 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4c9kk"] Oct 14 15:28:05 crc kubenswrapper[4945]: I1014 15:28:05.390276 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4c9kk" podUID="ac0ee420-6ad5-41cd-aad0-64777519ecda" containerName="registry-server" containerID="cri-o://5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd" gracePeriod=2 Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.323430 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.403925 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9cs8\" (UniqueName: \"kubernetes.io/projected/ac0ee420-6ad5-41cd-aad0-64777519ecda-kube-api-access-t9cs8\") pod \"ac0ee420-6ad5-41cd-aad0-64777519ecda\" (UID: \"ac0ee420-6ad5-41cd-aad0-64777519ecda\") " Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.403988 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac0ee420-6ad5-41cd-aad0-64777519ecda-utilities\") pod \"ac0ee420-6ad5-41cd-aad0-64777519ecda\" (UID: \"ac0ee420-6ad5-41cd-aad0-64777519ecda\") " Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.404085 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac0ee420-6ad5-41cd-aad0-64777519ecda-catalog-content\") pod \"ac0ee420-6ad5-41cd-aad0-64777519ecda\" (UID: \"ac0ee420-6ad5-41cd-aad0-64777519ecda\") " Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.405020 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac0ee420-6ad5-41cd-aad0-64777519ecda-utilities" (OuterVolumeSpecName: "utilities") pod "ac0ee420-6ad5-41cd-aad0-64777519ecda" (UID: "ac0ee420-6ad5-41cd-aad0-64777519ecda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.409085 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac0ee420-6ad5-41cd-aad0-64777519ecda-kube-api-access-t9cs8" (OuterVolumeSpecName: "kube-api-access-t9cs8") pod "ac0ee420-6ad5-41cd-aad0-64777519ecda" (UID: "ac0ee420-6ad5-41cd-aad0-64777519ecda"). InnerVolumeSpecName "kube-api-access-t9cs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.453671 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac0ee420-6ad5-41cd-aad0-64777519ecda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac0ee420-6ad5-41cd-aad0-64777519ecda" (UID: "ac0ee420-6ad5-41cd-aad0-64777519ecda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.505769 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac0ee420-6ad5-41cd-aad0-64777519ecda-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.505818 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9cs8\" (UniqueName: \"kubernetes.io/projected/ac0ee420-6ad5-41cd-aad0-64777519ecda-kube-api-access-t9cs8\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.505830 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac0ee420-6ad5-41cd-aad0-64777519ecda-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.717198 4945 generic.go:334] "Generic (PLEG): container finished" podID="ac0ee420-6ad5-41cd-aad0-64777519ecda" containerID="5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd" exitCode=0 Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.717250 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4c9kk" event={"ID":"ac0ee420-6ad5-41cd-aad0-64777519ecda","Type":"ContainerDied","Data":"5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd"} Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.717287 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4c9kk" event={"ID":"ac0ee420-6ad5-41cd-aad0-64777519ecda","Type":"ContainerDied","Data":"8c645b2a26c4623db67f81132c8286733aaccdddbdcb60387b2f3e98df2e4ea5"} Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.717335 4945 scope.go:117] "RemoveContainer" containerID="5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.717339 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4c9kk" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.750702 4945 scope.go:117] "RemoveContainer" containerID="7c7085ab41f62433d1cecdc71acdd037f65c89b8bb46b5924169fd3bc8705adb" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.755918 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4c9kk"] Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.771526 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4c9kk"] Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.772176 4945 scope.go:117] "RemoveContainer" containerID="cd144878953a03d1d49d22bdc42bc6546f92e47729884983d06b0fc8f1f9ae45" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.791434 4945 scope.go:117] "RemoveContainer" containerID="5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd" Oct 14 15:28:06 crc kubenswrapper[4945]: E1014 15:28:06.792105 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd\": container with ID starting with 5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd not found: ID does not exist" containerID="5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.792164 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd"} err="failed to get container status \"5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd\": rpc error: code = NotFound desc = could not find container \"5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd\": container with ID starting with 5538c4f7605dde6ccafd088e152733a93029995b55fcadd645905adff25c87fd not found: ID does not exist" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.792198 4945 scope.go:117] "RemoveContainer" containerID="7c7085ab41f62433d1cecdc71acdd037f65c89b8bb46b5924169fd3bc8705adb" Oct 14 15:28:06 crc kubenswrapper[4945]: E1014 15:28:06.792664 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c7085ab41f62433d1cecdc71acdd037f65c89b8bb46b5924169fd3bc8705adb\": container with ID starting with 7c7085ab41f62433d1cecdc71acdd037f65c89b8bb46b5924169fd3bc8705adb not found: ID does not exist" containerID="7c7085ab41f62433d1cecdc71acdd037f65c89b8bb46b5924169fd3bc8705adb" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.792711 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c7085ab41f62433d1cecdc71acdd037f65c89b8bb46b5924169fd3bc8705adb"} err="failed to get container status \"7c7085ab41f62433d1cecdc71acdd037f65c89b8bb46b5924169fd3bc8705adb\": rpc error: code = NotFound desc = could not find container \"7c7085ab41f62433d1cecdc71acdd037f65c89b8bb46b5924169fd3bc8705adb\": container with ID starting with 7c7085ab41f62433d1cecdc71acdd037f65c89b8bb46b5924169fd3bc8705adb not found: ID does not exist" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.792739 4945 scope.go:117] "RemoveContainer" containerID="cd144878953a03d1d49d22bdc42bc6546f92e47729884983d06b0fc8f1f9ae45" Oct 14 15:28:06 crc kubenswrapper[4945]: E1014 15:28:06.793265 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd144878953a03d1d49d22bdc42bc6546f92e47729884983d06b0fc8f1f9ae45\": container with ID starting with cd144878953a03d1d49d22bdc42bc6546f92e47729884983d06b0fc8f1f9ae45 not found: ID does not exist" containerID="cd144878953a03d1d49d22bdc42bc6546f92e47729884983d06b0fc8f1f9ae45" Oct 14 15:28:06 crc kubenswrapper[4945]: I1014 15:28:06.793294 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd144878953a03d1d49d22bdc42bc6546f92e47729884983d06b0fc8f1f9ae45"} err="failed to get container status \"cd144878953a03d1d49d22bdc42bc6546f92e47729884983d06b0fc8f1f9ae45\": rpc error: code = NotFound desc = could not find container \"cd144878953a03d1d49d22bdc42bc6546f92e47729884983d06b0fc8f1f9ae45\": container with ID starting with cd144878953a03d1d49d22bdc42bc6546f92e47729884983d06b0fc8f1f9ae45 not found: ID does not exist" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.605629 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9dd52"] Oct 14 15:28:07 crc kubenswrapper[4945]: E1014 15:28:07.606231 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac0ee420-6ad5-41cd-aad0-64777519ecda" containerName="extract-content" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.606282 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac0ee420-6ad5-41cd-aad0-64777519ecda" containerName="extract-content" Oct 14 15:28:07 crc kubenswrapper[4945]: E1014 15:28:07.606316 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac0ee420-6ad5-41cd-aad0-64777519ecda" containerName="registry-server" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.606328 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac0ee420-6ad5-41cd-aad0-64777519ecda" containerName="registry-server" Oct 14 15:28:07 crc kubenswrapper[4945]: E1014 15:28:07.606359 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac0ee420-6ad5-41cd-aad0-64777519ecda" containerName="extract-utilities" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.606371 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac0ee420-6ad5-41cd-aad0-64777519ecda" containerName="extract-utilities" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.606642 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac0ee420-6ad5-41cd-aad0-64777519ecda" containerName="registry-server" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.608527 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.624107 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9dd52"] Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.722380 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h5nz\" (UniqueName: \"kubernetes.io/projected/d554d7c4-c5b2-457d-99f3-6496ed9de514-kube-api-access-7h5nz\") pod \"redhat-operators-9dd52\" (UID: \"d554d7c4-c5b2-457d-99f3-6496ed9de514\") " pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.722556 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d554d7c4-c5b2-457d-99f3-6496ed9de514-utilities\") pod \"redhat-operators-9dd52\" (UID: \"d554d7c4-c5b2-457d-99f3-6496ed9de514\") " pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.722611 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d554d7c4-c5b2-457d-99f3-6496ed9de514-catalog-content\") pod \"redhat-operators-9dd52\" (UID: \"d554d7c4-c5b2-457d-99f3-6496ed9de514\") " pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.823774 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h5nz\" (UniqueName: \"kubernetes.io/projected/d554d7c4-c5b2-457d-99f3-6496ed9de514-kube-api-access-7h5nz\") pod \"redhat-operators-9dd52\" (UID: \"d554d7c4-c5b2-457d-99f3-6496ed9de514\") " pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.823859 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d554d7c4-c5b2-457d-99f3-6496ed9de514-utilities\") pod \"redhat-operators-9dd52\" (UID: \"d554d7c4-c5b2-457d-99f3-6496ed9de514\") " pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.823901 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d554d7c4-c5b2-457d-99f3-6496ed9de514-catalog-content\") pod \"redhat-operators-9dd52\" (UID: \"d554d7c4-c5b2-457d-99f3-6496ed9de514\") " pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.824300 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d554d7c4-c5b2-457d-99f3-6496ed9de514-utilities\") pod \"redhat-operators-9dd52\" (UID: \"d554d7c4-c5b2-457d-99f3-6496ed9de514\") " pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.824388 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d554d7c4-c5b2-457d-99f3-6496ed9de514-catalog-content\") pod \"redhat-operators-9dd52\" (UID: \"d554d7c4-c5b2-457d-99f3-6496ed9de514\") " pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.840177 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h5nz\" (UniqueName: \"kubernetes.io/projected/d554d7c4-c5b2-457d-99f3-6496ed9de514-kube-api-access-7h5nz\") pod \"redhat-operators-9dd52\" (UID: \"d554d7c4-c5b2-457d-99f3-6496ed9de514\") " pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:07 crc kubenswrapper[4945]: I1014 15:28:07.944454 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:08 crc kubenswrapper[4945]: I1014 15:28:08.353397 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9dd52"] Oct 14 15:28:08 crc kubenswrapper[4945]: I1014 15:28:08.571902 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-f9h5q" Oct 14 15:28:08 crc kubenswrapper[4945]: I1014 15:28:08.572232 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-f9h5q" Oct 14 15:28:08 crc kubenswrapper[4945]: I1014 15:28:08.603410 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-f9h5q" Oct 14 15:28:08 crc kubenswrapper[4945]: I1014 15:28:08.731219 4945 generic.go:334] "Generic (PLEG): container finished" podID="d554d7c4-c5b2-457d-99f3-6496ed9de514" containerID="123042d8cbf767325119ba283aa4cd5c99d1f0ccd646c9fcaf93aeb8ac02612d" exitCode=0 Oct 14 15:28:08 crc kubenswrapper[4945]: I1014 15:28:08.731372 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dd52" event={"ID":"d554d7c4-c5b2-457d-99f3-6496ed9de514","Type":"ContainerDied","Data":"123042d8cbf767325119ba283aa4cd5c99d1f0ccd646c9fcaf93aeb8ac02612d"} Oct 14 15:28:08 crc kubenswrapper[4945]: I1014 15:28:08.731444 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dd52" event={"ID":"d554d7c4-c5b2-457d-99f3-6496ed9de514","Type":"ContainerStarted","Data":"211db9eef81e6dae55fe4c6c0e7dd163882fd883c2796348980c6075417d16b2"} Oct 14 15:28:08 crc kubenswrapper[4945]: I1014 15:28:08.758521 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-f9h5q" Oct 14 15:28:08 crc kubenswrapper[4945]: I1014 15:28:08.771276 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac0ee420-6ad5-41cd-aad0-64777519ecda" path="/var/lib/kubelet/pods/ac0ee420-6ad5-41cd-aad0-64777519ecda/volumes" Oct 14 15:28:09 crc kubenswrapper[4945]: I1014 15:28:09.739096 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dd52" event={"ID":"d554d7c4-c5b2-457d-99f3-6496ed9de514","Type":"ContainerStarted","Data":"8a516813f027de3687470f9872c3a349b4908b9ce9f94328a540cc9069396ed0"} Oct 14 15:28:10 crc kubenswrapper[4945]: I1014 15:28:10.748141 4945 generic.go:334] "Generic (PLEG): container finished" podID="d554d7c4-c5b2-457d-99f3-6496ed9de514" containerID="8a516813f027de3687470f9872c3a349b4908b9ce9f94328a540cc9069396ed0" exitCode=0 Oct 14 15:28:10 crc kubenswrapper[4945]: I1014 15:28:10.748188 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dd52" event={"ID":"d554d7c4-c5b2-457d-99f3-6496ed9de514","Type":"ContainerDied","Data":"8a516813f027de3687470f9872c3a349b4908b9ce9f94328a540cc9069396ed0"} Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.233424 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md"] Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.235848 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.239947 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-mg4km" Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.244683 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md"] Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.270011 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d84b1643-7a42-4df5-ad42-d31cf321ce9d-bundle\") pod \"e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md\" (UID: \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\") " pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.270070 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d84b1643-7a42-4df5-ad42-d31cf321ce9d-util\") pod \"e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md\" (UID: \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\") " pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.270150 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wszm\" (UniqueName: \"kubernetes.io/projected/d84b1643-7a42-4df5-ad42-d31cf321ce9d-kube-api-access-5wszm\") pod \"e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md\" (UID: \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\") " pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.371669 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d84b1643-7a42-4df5-ad42-d31cf321ce9d-util\") pod \"e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md\" (UID: \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\") " pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.371795 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wszm\" (UniqueName: \"kubernetes.io/projected/d84b1643-7a42-4df5-ad42-d31cf321ce9d-kube-api-access-5wszm\") pod \"e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md\" (UID: \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\") " pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.371842 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d84b1643-7a42-4df5-ad42-d31cf321ce9d-bundle\") pod \"e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md\" (UID: \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\") " pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.372370 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d84b1643-7a42-4df5-ad42-d31cf321ce9d-util\") pod \"e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md\" (UID: \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\") " pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.372392 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d84b1643-7a42-4df5-ad42-d31cf321ce9d-bundle\") pod \"e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md\" (UID: \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\") " pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.394023 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wszm\" (UniqueName: \"kubernetes.io/projected/d84b1643-7a42-4df5-ad42-d31cf321ce9d-kube-api-access-5wszm\") pod \"e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md\" (UID: \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\") " pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.554194 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.763350 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dd52" event={"ID":"d554d7c4-c5b2-457d-99f3-6496ed9de514","Type":"ContainerStarted","Data":"fc901807255ba5875ad8bdd62c765957d4044c98bb947771ddf6afa1e3d05169"} Oct 14 15:28:11 crc kubenswrapper[4945]: I1014 15:28:11.779239 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md"] Oct 14 15:28:12 crc kubenswrapper[4945]: I1014 15:28:12.771338 4945 generic.go:334] "Generic (PLEG): container finished" podID="d84b1643-7a42-4df5-ad42-d31cf321ce9d" containerID="5dff6ae624f13011f6f09ffa386f43579d62dc5dd072cf08ed45d3a325bcd97c" exitCode=0 Oct 14 15:28:12 crc kubenswrapper[4945]: I1014 15:28:12.774932 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" event={"ID":"d84b1643-7a42-4df5-ad42-d31cf321ce9d","Type":"ContainerDied","Data":"5dff6ae624f13011f6f09ffa386f43579d62dc5dd072cf08ed45d3a325bcd97c"} Oct 14 15:28:12 crc kubenswrapper[4945]: I1014 15:28:12.774975 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" event={"ID":"d84b1643-7a42-4df5-ad42-d31cf321ce9d","Type":"ContainerStarted","Data":"ec7edfac11102d1b03748f2c1aced64c73e5407e8ee270a1edb83fa98721b2ee"} Oct 14 15:28:12 crc kubenswrapper[4945]: I1014 15:28:12.798985 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9dd52" podStartSLOduration=3.286560649 podStartE2EDuration="5.79896197s" podCreationTimestamp="2025-10-14 15:28:07 +0000 UTC" firstStartedPulling="2025-10-14 15:28:08.732793459 +0000 UTC m=+818.716841827" lastFinishedPulling="2025-10-14 15:28:11.24519478 +0000 UTC m=+821.229243148" observedRunningTime="2025-10-14 15:28:11.78617586 +0000 UTC m=+821.770224238" watchObservedRunningTime="2025-10-14 15:28:12.79896197 +0000 UTC m=+822.783010358" Oct 14 15:28:13 crc kubenswrapper[4945]: I1014 15:28:13.784968 4945 generic.go:334] "Generic (PLEG): container finished" podID="d84b1643-7a42-4df5-ad42-d31cf321ce9d" containerID="ddde0f1b8822e50da7c47ad261eeaac59acf24039cc6f20d7f847eb22cb638b7" exitCode=0 Oct 14 15:28:13 crc kubenswrapper[4945]: I1014 15:28:13.785029 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" event={"ID":"d84b1643-7a42-4df5-ad42-d31cf321ce9d","Type":"ContainerDied","Data":"ddde0f1b8822e50da7c47ad261eeaac59acf24039cc6f20d7f847eb22cb638b7"} Oct 14 15:28:14 crc kubenswrapper[4945]: I1014 15:28:14.792400 4945 generic.go:334] "Generic (PLEG): container finished" podID="d84b1643-7a42-4df5-ad42-d31cf321ce9d" containerID="d1f3b0a364c3a190066cfffce345c57f94ec1f7839edff147ae6ad3e55ab24e1" exitCode=0 Oct 14 15:28:14 crc kubenswrapper[4945]: I1014 15:28:14.792439 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" event={"ID":"d84b1643-7a42-4df5-ad42-d31cf321ce9d","Type":"ContainerDied","Data":"d1f3b0a364c3a190066cfffce345c57f94ec1f7839edff147ae6ad3e55ab24e1"} Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.399031 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vchw9"] Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.400637 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.419298 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vchw9"] Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.429712 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f9c79c-b544-43a4-a794-21293b2de0f9-catalog-content\") pod \"certified-operators-vchw9\" (UID: \"02f9c79c-b544-43a4-a794-21293b2de0f9\") " pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.429814 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f9c79c-b544-43a4-a794-21293b2de0f9-utilities\") pod \"certified-operators-vchw9\" (UID: \"02f9c79c-b544-43a4-a794-21293b2de0f9\") " pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.429842 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5qc2\" (UniqueName: \"kubernetes.io/projected/02f9c79c-b544-43a4-a794-21293b2de0f9-kube-api-access-g5qc2\") pod \"certified-operators-vchw9\" (UID: \"02f9c79c-b544-43a4-a794-21293b2de0f9\") " pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.531161 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f9c79c-b544-43a4-a794-21293b2de0f9-catalog-content\") pod \"certified-operators-vchw9\" (UID: \"02f9c79c-b544-43a4-a794-21293b2de0f9\") " pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.531246 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f9c79c-b544-43a4-a794-21293b2de0f9-utilities\") pod \"certified-operators-vchw9\" (UID: \"02f9c79c-b544-43a4-a794-21293b2de0f9\") " pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.531265 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5qc2\" (UniqueName: \"kubernetes.io/projected/02f9c79c-b544-43a4-a794-21293b2de0f9-kube-api-access-g5qc2\") pod \"certified-operators-vchw9\" (UID: \"02f9c79c-b544-43a4-a794-21293b2de0f9\") " pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.531744 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f9c79c-b544-43a4-a794-21293b2de0f9-catalog-content\") pod \"certified-operators-vchw9\" (UID: \"02f9c79c-b544-43a4-a794-21293b2de0f9\") " pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.531919 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f9c79c-b544-43a4-a794-21293b2de0f9-utilities\") pod \"certified-operators-vchw9\" (UID: \"02f9c79c-b544-43a4-a794-21293b2de0f9\") " pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.550585 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5qc2\" (UniqueName: \"kubernetes.io/projected/02f9c79c-b544-43a4-a794-21293b2de0f9-kube-api-access-g5qc2\") pod \"certified-operators-vchw9\" (UID: \"02f9c79c-b544-43a4-a794-21293b2de0f9\") " pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:15 crc kubenswrapper[4945]: I1014 15:28:15.722200 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.152305 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.192193 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vchw9"] Oct 14 15:28:16 crc kubenswrapper[4945]: W1014 15:28:16.206514 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02f9c79c_b544_43a4_a794_21293b2de0f9.slice/crio-5fcaa98a4dd3dcf368fbb894ec0a41524152097c2636ffbc48ba43f9a31a7cad WatchSource:0}: Error finding container 5fcaa98a4dd3dcf368fbb894ec0a41524152097c2636ffbc48ba43f9a31a7cad: Status 404 returned error can't find the container with id 5fcaa98a4dd3dcf368fbb894ec0a41524152097c2636ffbc48ba43f9a31a7cad Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.239477 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d84b1643-7a42-4df5-ad42-d31cf321ce9d-bundle\") pod \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\" (UID: \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\") " Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.239524 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wszm\" (UniqueName: \"kubernetes.io/projected/d84b1643-7a42-4df5-ad42-d31cf321ce9d-kube-api-access-5wszm\") pod \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\" (UID: \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\") " Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.239594 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d84b1643-7a42-4df5-ad42-d31cf321ce9d-util\") pod \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\" (UID: \"d84b1643-7a42-4df5-ad42-d31cf321ce9d\") " Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.240626 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d84b1643-7a42-4df5-ad42-d31cf321ce9d-bundle" (OuterVolumeSpecName: "bundle") pod "d84b1643-7a42-4df5-ad42-d31cf321ce9d" (UID: "d84b1643-7a42-4df5-ad42-d31cf321ce9d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.245753 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d84b1643-7a42-4df5-ad42-d31cf321ce9d-kube-api-access-5wszm" (OuterVolumeSpecName: "kube-api-access-5wszm") pod "d84b1643-7a42-4df5-ad42-d31cf321ce9d" (UID: "d84b1643-7a42-4df5-ad42-d31cf321ce9d"). InnerVolumeSpecName "kube-api-access-5wszm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.256858 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d84b1643-7a42-4df5-ad42-d31cf321ce9d-util" (OuterVolumeSpecName: "util") pod "d84b1643-7a42-4df5-ad42-d31cf321ce9d" (UID: "d84b1643-7a42-4df5-ad42-d31cf321ce9d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.341038 4945 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d84b1643-7a42-4df5-ad42-d31cf321ce9d-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.341078 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wszm\" (UniqueName: \"kubernetes.io/projected/d84b1643-7a42-4df5-ad42-d31cf321ce9d-kube-api-access-5wszm\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.341092 4945 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d84b1643-7a42-4df5-ad42-d31cf321ce9d-util\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.795431 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.795711 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.805685 4945 generic.go:334] "Generic (PLEG): container finished" podID="02f9c79c-b544-43a4-a794-21293b2de0f9" containerID="9bbc3bc9d108614a1a523cd73a2af6df5f923fc02e52e734dc80ad7a56635fff" exitCode=0 Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.805750 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vchw9" event={"ID":"02f9c79c-b544-43a4-a794-21293b2de0f9","Type":"ContainerDied","Data":"9bbc3bc9d108614a1a523cd73a2af6df5f923fc02e52e734dc80ad7a56635fff"} Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.805776 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vchw9" event={"ID":"02f9c79c-b544-43a4-a794-21293b2de0f9","Type":"ContainerStarted","Data":"5fcaa98a4dd3dcf368fbb894ec0a41524152097c2636ffbc48ba43f9a31a7cad"} Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.808608 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" event={"ID":"d84b1643-7a42-4df5-ad42-d31cf321ce9d","Type":"ContainerDied","Data":"ec7edfac11102d1b03748f2c1aced64c73e5407e8ee270a1edb83fa98721b2ee"} Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.808633 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec7edfac11102d1b03748f2c1aced64c73e5407e8ee270a1edb83fa98721b2ee" Oct 14 15:28:16 crc kubenswrapper[4945]: I1014 15:28:16.808750 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md" Oct 14 15:28:17 crc kubenswrapper[4945]: I1014 15:28:17.817051 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vchw9" event={"ID":"02f9c79c-b544-43a4-a794-21293b2de0f9","Type":"ContainerStarted","Data":"82dff883e6eca88859eae896506c89ca33a6c1fb150b43f3e95041698375cffc"} Oct 14 15:28:17 crc kubenswrapper[4945]: I1014 15:28:17.945286 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:17 crc kubenswrapper[4945]: I1014 15:28:17.945389 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:18 crc kubenswrapper[4945]: I1014 15:28:18.003577 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:18 crc kubenswrapper[4945]: I1014 15:28:18.829405 4945 generic.go:334] "Generic (PLEG): container finished" podID="02f9c79c-b544-43a4-a794-21293b2de0f9" containerID="82dff883e6eca88859eae896506c89ca33a6c1fb150b43f3e95041698375cffc" exitCode=0 Oct 14 15:28:18 crc kubenswrapper[4945]: I1014 15:28:18.829470 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vchw9" event={"ID":"02f9c79c-b544-43a4-a794-21293b2de0f9","Type":"ContainerDied","Data":"82dff883e6eca88859eae896506c89ca33a6c1fb150b43f3e95041698375cffc"} Oct 14 15:28:18 crc kubenswrapper[4945]: I1014 15:28:18.882497 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:19 crc kubenswrapper[4945]: I1014 15:28:19.838619 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vchw9" event={"ID":"02f9c79c-b544-43a4-a794-21293b2de0f9","Type":"ContainerStarted","Data":"075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19"} Oct 14 15:28:19 crc kubenswrapper[4945]: I1014 15:28:19.866313 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vchw9" podStartSLOduration=2.257605047 podStartE2EDuration="4.86629623s" podCreationTimestamp="2025-10-14 15:28:15 +0000 UTC" firstStartedPulling="2025-10-14 15:28:16.808510539 +0000 UTC m=+826.792558917" lastFinishedPulling="2025-10-14 15:28:19.417201722 +0000 UTC m=+829.401250100" observedRunningTime="2025-10-14 15:28:19.863500448 +0000 UTC m=+829.847548836" watchObservedRunningTime="2025-10-14 15:28:19.86629623 +0000 UTC m=+829.850344598" Oct 14 15:28:20 crc kubenswrapper[4945]: I1014 15:28:20.194965 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9dd52"] Oct 14 15:28:20 crc kubenswrapper[4945]: I1014 15:28:20.843761 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9dd52" podUID="d554d7c4-c5b2-457d-99f3-6496ed9de514" containerName="registry-server" containerID="cri-o://fc901807255ba5875ad8bdd62c765957d4044c98bb947771ddf6afa1e3d05169" gracePeriod=2 Oct 14 15:28:21 crc kubenswrapper[4945]: I1014 15:28:21.850828 4945 generic.go:334] "Generic (PLEG): container finished" podID="d554d7c4-c5b2-457d-99f3-6496ed9de514" containerID="fc901807255ba5875ad8bdd62c765957d4044c98bb947771ddf6afa1e3d05169" exitCode=0 Oct 14 15:28:21 crc kubenswrapper[4945]: I1014 15:28:21.850930 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dd52" event={"ID":"d554d7c4-c5b2-457d-99f3-6496ed9de514","Type":"ContainerDied","Data":"fc901807255ba5875ad8bdd62c765957d4044c98bb947771ddf6afa1e3d05169"} Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.414857 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.546291 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d554d7c4-c5b2-457d-99f3-6496ed9de514-utilities\") pod \"d554d7c4-c5b2-457d-99f3-6496ed9de514\" (UID: \"d554d7c4-c5b2-457d-99f3-6496ed9de514\") " Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.546348 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h5nz\" (UniqueName: \"kubernetes.io/projected/d554d7c4-c5b2-457d-99f3-6496ed9de514-kube-api-access-7h5nz\") pod \"d554d7c4-c5b2-457d-99f3-6496ed9de514\" (UID: \"d554d7c4-c5b2-457d-99f3-6496ed9de514\") " Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.546436 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d554d7c4-c5b2-457d-99f3-6496ed9de514-catalog-content\") pod \"d554d7c4-c5b2-457d-99f3-6496ed9de514\" (UID: \"d554d7c4-c5b2-457d-99f3-6496ed9de514\") " Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.547207 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d554d7c4-c5b2-457d-99f3-6496ed9de514-utilities" (OuterVolumeSpecName: "utilities") pod "d554d7c4-c5b2-457d-99f3-6496ed9de514" (UID: "d554d7c4-c5b2-457d-99f3-6496ed9de514"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.552178 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d554d7c4-c5b2-457d-99f3-6496ed9de514-kube-api-access-7h5nz" (OuterVolumeSpecName: "kube-api-access-7h5nz") pod "d554d7c4-c5b2-457d-99f3-6496ed9de514" (UID: "d554d7c4-c5b2-457d-99f3-6496ed9de514"). InnerVolumeSpecName "kube-api-access-7h5nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.651150 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d554d7c4-c5b2-457d-99f3-6496ed9de514-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.651193 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h5nz\" (UniqueName: \"kubernetes.io/projected/d554d7c4-c5b2-457d-99f3-6496ed9de514-kube-api-access-7h5nz\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.658716 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d554d7c4-c5b2-457d-99f3-6496ed9de514-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d554d7c4-c5b2-457d-99f3-6496ed9de514" (UID: "d554d7c4-c5b2-457d-99f3-6496ed9de514"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.752680 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d554d7c4-c5b2-457d-99f3-6496ed9de514-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.859959 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dd52" event={"ID":"d554d7c4-c5b2-457d-99f3-6496ed9de514","Type":"ContainerDied","Data":"211db9eef81e6dae55fe4c6c0e7dd163882fd883c2796348980c6075417d16b2"} Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.860035 4945 scope.go:117] "RemoveContainer" containerID="fc901807255ba5875ad8bdd62c765957d4044c98bb947771ddf6afa1e3d05169" Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.860039 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dd52" Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.875912 4945 scope.go:117] "RemoveContainer" containerID="8a516813f027de3687470f9872c3a349b4908b9ce9f94328a540cc9069396ed0" Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.879363 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9dd52"] Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.884129 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9dd52"] Oct 14 15:28:22 crc kubenswrapper[4945]: I1014 15:28:22.892449 4945 scope.go:117] "RemoveContainer" containerID="123042d8cbf767325119ba283aa4cd5c99d1f0ccd646c9fcaf93aeb8ac02612d" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.251385 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6"] Oct 14 15:28:23 crc kubenswrapper[4945]: E1014 15:28:23.251628 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d554d7c4-c5b2-457d-99f3-6496ed9de514" containerName="extract-utilities" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.251646 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d554d7c4-c5b2-457d-99f3-6496ed9de514" containerName="extract-utilities" Oct 14 15:28:23 crc kubenswrapper[4945]: E1014 15:28:23.251656 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84b1643-7a42-4df5-ad42-d31cf321ce9d" containerName="pull" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.251664 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84b1643-7a42-4df5-ad42-d31cf321ce9d" containerName="pull" Oct 14 15:28:23 crc kubenswrapper[4945]: E1014 15:28:23.251673 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84b1643-7a42-4df5-ad42-d31cf321ce9d" containerName="extract" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.251679 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84b1643-7a42-4df5-ad42-d31cf321ce9d" containerName="extract" Oct 14 15:28:23 crc kubenswrapper[4945]: E1014 15:28:23.251691 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d554d7c4-c5b2-457d-99f3-6496ed9de514" containerName="extract-content" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.251698 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d554d7c4-c5b2-457d-99f3-6496ed9de514" containerName="extract-content" Oct 14 15:28:23 crc kubenswrapper[4945]: E1014 15:28:23.251715 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d554d7c4-c5b2-457d-99f3-6496ed9de514" containerName="registry-server" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.251720 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d554d7c4-c5b2-457d-99f3-6496ed9de514" containerName="registry-server" Oct 14 15:28:23 crc kubenswrapper[4945]: E1014 15:28:23.251727 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84b1643-7a42-4df5-ad42-d31cf321ce9d" containerName="util" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.251733 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84b1643-7a42-4df5-ad42-d31cf321ce9d" containerName="util" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.251844 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d84b1643-7a42-4df5-ad42-d31cf321ce9d" containerName="extract" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.251859 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d554d7c4-c5b2-457d-99f3-6496ed9de514" containerName="registry-server" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.252445 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.254386 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-5dgqb" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.277816 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6"] Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.361905 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8pmh\" (UniqueName: \"kubernetes.io/projected/795c3628-2ec0-4846-b345-3019636282e9-kube-api-access-x8pmh\") pod \"openstack-operator-controller-operator-674f597f57-cqpn6\" (UID: \"795c3628-2ec0-4846-b345-3019636282e9\") " pod="openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.463766 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8pmh\" (UniqueName: \"kubernetes.io/projected/795c3628-2ec0-4846-b345-3019636282e9-kube-api-access-x8pmh\") pod \"openstack-operator-controller-operator-674f597f57-cqpn6\" (UID: \"795c3628-2ec0-4846-b345-3019636282e9\") " pod="openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.493433 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8pmh\" (UniqueName: \"kubernetes.io/projected/795c3628-2ec0-4846-b345-3019636282e9-kube-api-access-x8pmh\") pod \"openstack-operator-controller-operator-674f597f57-cqpn6\" (UID: \"795c3628-2ec0-4846-b345-3019636282e9\") " pod="openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.574280 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.798936 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mpc78"] Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.800430 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.811813 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpc78"] Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.869685 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a990723-8b29-4962-8861-f692e2da3bcd-catalog-content\") pod \"redhat-marketplace-mpc78\" (UID: \"7a990723-8b29-4962-8861-f692e2da3bcd\") " pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.869736 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a990723-8b29-4962-8861-f692e2da3bcd-utilities\") pod \"redhat-marketplace-mpc78\" (UID: \"7a990723-8b29-4962-8861-f692e2da3bcd\") " pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.869758 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjhdh\" (UniqueName: \"kubernetes.io/projected/7a990723-8b29-4962-8861-f692e2da3bcd-kube-api-access-vjhdh\") pod \"redhat-marketplace-mpc78\" (UID: \"7a990723-8b29-4962-8861-f692e2da3bcd\") " pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.877439 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6"] Oct 14 15:28:23 crc kubenswrapper[4945]: W1014 15:28:23.892433 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod795c3628_2ec0_4846_b345_3019636282e9.slice/crio-53f28b02f5ed80d6eded6ef3c2fb4c041374009520c4aac7d59e8e48e798a24d WatchSource:0}: Error finding container 53f28b02f5ed80d6eded6ef3c2fb4c041374009520c4aac7d59e8e48e798a24d: Status 404 returned error can't find the container with id 53f28b02f5ed80d6eded6ef3c2fb4c041374009520c4aac7d59e8e48e798a24d Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.970449 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a990723-8b29-4962-8861-f692e2da3bcd-catalog-content\") pod \"redhat-marketplace-mpc78\" (UID: \"7a990723-8b29-4962-8861-f692e2da3bcd\") " pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.970507 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a990723-8b29-4962-8861-f692e2da3bcd-utilities\") pod \"redhat-marketplace-mpc78\" (UID: \"7a990723-8b29-4962-8861-f692e2da3bcd\") " pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.970538 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjhdh\" (UniqueName: \"kubernetes.io/projected/7a990723-8b29-4962-8861-f692e2da3bcd-kube-api-access-vjhdh\") pod \"redhat-marketplace-mpc78\" (UID: \"7a990723-8b29-4962-8861-f692e2da3bcd\") " pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.971218 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a990723-8b29-4962-8861-f692e2da3bcd-utilities\") pod \"redhat-marketplace-mpc78\" (UID: \"7a990723-8b29-4962-8861-f692e2da3bcd\") " pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.971408 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a990723-8b29-4962-8861-f692e2da3bcd-catalog-content\") pod \"redhat-marketplace-mpc78\" (UID: \"7a990723-8b29-4962-8861-f692e2da3bcd\") " pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:23 crc kubenswrapper[4945]: I1014 15:28:23.988253 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjhdh\" (UniqueName: \"kubernetes.io/projected/7a990723-8b29-4962-8861-f692e2da3bcd-kube-api-access-vjhdh\") pod \"redhat-marketplace-mpc78\" (UID: \"7a990723-8b29-4962-8861-f692e2da3bcd\") " pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:24 crc kubenswrapper[4945]: I1014 15:28:24.117633 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:24 crc kubenswrapper[4945]: I1014 15:28:24.415664 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpc78"] Oct 14 15:28:24 crc kubenswrapper[4945]: I1014 15:28:24.770963 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d554d7c4-c5b2-457d-99f3-6496ed9de514" path="/var/lib/kubelet/pods/d554d7c4-c5b2-457d-99f3-6496ed9de514/volumes" Oct 14 15:28:24 crc kubenswrapper[4945]: I1014 15:28:24.880259 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6" event={"ID":"795c3628-2ec0-4846-b345-3019636282e9","Type":"ContainerStarted","Data":"53f28b02f5ed80d6eded6ef3c2fb4c041374009520c4aac7d59e8e48e798a24d"} Oct 14 15:28:24 crc kubenswrapper[4945]: I1014 15:28:24.884432 4945 generic.go:334] "Generic (PLEG): container finished" podID="7a990723-8b29-4962-8861-f692e2da3bcd" containerID="a192378b08532bfd30ad108bd03a8303e0c3dbec56697039c4633c51510a3dc9" exitCode=0 Oct 14 15:28:24 crc kubenswrapper[4945]: I1014 15:28:24.884480 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpc78" event={"ID":"7a990723-8b29-4962-8861-f692e2da3bcd","Type":"ContainerDied","Data":"a192378b08532bfd30ad108bd03a8303e0c3dbec56697039c4633c51510a3dc9"} Oct 14 15:28:24 crc kubenswrapper[4945]: I1014 15:28:24.884512 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpc78" event={"ID":"7a990723-8b29-4962-8861-f692e2da3bcd","Type":"ContainerStarted","Data":"326cd262292a88e8830c63be0dcee8cedcdf56917bb695ac46dadbed41c0197d"} Oct 14 15:28:25 crc kubenswrapper[4945]: I1014 15:28:25.723129 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:25 crc kubenswrapper[4945]: I1014 15:28:25.723183 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:25 crc kubenswrapper[4945]: I1014 15:28:25.763255 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:25 crc kubenswrapper[4945]: I1014 15:28:25.935735 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:27 crc kubenswrapper[4945]: I1014 15:28:27.589949 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vchw9"] Oct 14 15:28:27 crc kubenswrapper[4945]: I1014 15:28:27.903685 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vchw9" podUID="02f9c79c-b544-43a4-a794-21293b2de0f9" containerName="registry-server" containerID="cri-o://075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19" gracePeriod=2 Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.756310 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.854712 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f9c79c-b544-43a4-a794-21293b2de0f9-utilities\") pod \"02f9c79c-b544-43a4-a794-21293b2de0f9\" (UID: \"02f9c79c-b544-43a4-a794-21293b2de0f9\") " Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.854813 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5qc2\" (UniqueName: \"kubernetes.io/projected/02f9c79c-b544-43a4-a794-21293b2de0f9-kube-api-access-g5qc2\") pod \"02f9c79c-b544-43a4-a794-21293b2de0f9\" (UID: \"02f9c79c-b544-43a4-a794-21293b2de0f9\") " Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.854932 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f9c79c-b544-43a4-a794-21293b2de0f9-catalog-content\") pod \"02f9c79c-b544-43a4-a794-21293b2de0f9\" (UID: \"02f9c79c-b544-43a4-a794-21293b2de0f9\") " Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.856339 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02f9c79c-b544-43a4-a794-21293b2de0f9-utilities" (OuterVolumeSpecName: "utilities") pod "02f9c79c-b544-43a4-a794-21293b2de0f9" (UID: "02f9c79c-b544-43a4-a794-21293b2de0f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.864169 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02f9c79c-b544-43a4-a794-21293b2de0f9-kube-api-access-g5qc2" (OuterVolumeSpecName: "kube-api-access-g5qc2") pod "02f9c79c-b544-43a4-a794-21293b2de0f9" (UID: "02f9c79c-b544-43a4-a794-21293b2de0f9"). InnerVolumeSpecName "kube-api-access-g5qc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.915506 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpc78" event={"ID":"7a990723-8b29-4962-8861-f692e2da3bcd","Type":"ContainerStarted","Data":"d775cbfc900ad9db579052a46a21f1849df1313e45345912b96b26c4468fe575"} Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.919650 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6" event={"ID":"795c3628-2ec0-4846-b345-3019636282e9","Type":"ContainerStarted","Data":"7ad6f51db7c83b1220d066a13e8a5499f91558ace08f6d0d87e9a469b7a7ecbb"} Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.923802 4945 generic.go:334] "Generic (PLEG): container finished" podID="02f9c79c-b544-43a4-a794-21293b2de0f9" containerID="075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19" exitCode=0 Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.923845 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vchw9" event={"ID":"02f9c79c-b544-43a4-a794-21293b2de0f9","Type":"ContainerDied","Data":"075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19"} Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.923886 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vchw9" event={"ID":"02f9c79c-b544-43a4-a794-21293b2de0f9","Type":"ContainerDied","Data":"5fcaa98a4dd3dcf368fbb894ec0a41524152097c2636ffbc48ba43f9a31a7cad"} Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.923911 4945 scope.go:117] "RemoveContainer" containerID="075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.924045 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vchw9" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.941919 4945 scope.go:117] "RemoveContainer" containerID="82dff883e6eca88859eae896506c89ca33a6c1fb150b43f3e95041698375cffc" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.957443 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f9c79c-b544-43a4-a794-21293b2de0f9-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.957490 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5qc2\" (UniqueName: \"kubernetes.io/projected/02f9c79c-b544-43a4-a794-21293b2de0f9-kube-api-access-g5qc2\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.964489 4945 scope.go:117] "RemoveContainer" containerID="9bbc3bc9d108614a1a523cd73a2af6df5f923fc02e52e734dc80ad7a56635fff" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.985027 4945 scope.go:117] "RemoveContainer" containerID="075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19" Oct 14 15:28:28 crc kubenswrapper[4945]: E1014 15:28:28.985810 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19\": container with ID starting with 075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19 not found: ID does not exist" containerID="075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.985978 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19"} err="failed to get container status \"075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19\": rpc error: code = NotFound desc = could not find container \"075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19\": container with ID starting with 075f977aa554c2823d95d2c9fcc212916407b5e1cb8fe7a7ffda14545c859f19 not found: ID does not exist" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.986045 4945 scope.go:117] "RemoveContainer" containerID="82dff883e6eca88859eae896506c89ca33a6c1fb150b43f3e95041698375cffc" Oct 14 15:28:28 crc kubenswrapper[4945]: E1014 15:28:28.986414 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82dff883e6eca88859eae896506c89ca33a6c1fb150b43f3e95041698375cffc\": container with ID starting with 82dff883e6eca88859eae896506c89ca33a6c1fb150b43f3e95041698375cffc not found: ID does not exist" containerID="82dff883e6eca88859eae896506c89ca33a6c1fb150b43f3e95041698375cffc" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.986464 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82dff883e6eca88859eae896506c89ca33a6c1fb150b43f3e95041698375cffc"} err="failed to get container status \"82dff883e6eca88859eae896506c89ca33a6c1fb150b43f3e95041698375cffc\": rpc error: code = NotFound desc = could not find container \"82dff883e6eca88859eae896506c89ca33a6c1fb150b43f3e95041698375cffc\": container with ID starting with 82dff883e6eca88859eae896506c89ca33a6c1fb150b43f3e95041698375cffc not found: ID does not exist" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.986481 4945 scope.go:117] "RemoveContainer" containerID="9bbc3bc9d108614a1a523cd73a2af6df5f923fc02e52e734dc80ad7a56635fff" Oct 14 15:28:28 crc kubenswrapper[4945]: E1014 15:28:28.986820 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bbc3bc9d108614a1a523cd73a2af6df5f923fc02e52e734dc80ad7a56635fff\": container with ID starting with 9bbc3bc9d108614a1a523cd73a2af6df5f923fc02e52e734dc80ad7a56635fff not found: ID does not exist" containerID="9bbc3bc9d108614a1a523cd73a2af6df5f923fc02e52e734dc80ad7a56635fff" Oct 14 15:28:28 crc kubenswrapper[4945]: I1014 15:28:28.986899 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bbc3bc9d108614a1a523cd73a2af6df5f923fc02e52e734dc80ad7a56635fff"} err="failed to get container status \"9bbc3bc9d108614a1a523cd73a2af6df5f923fc02e52e734dc80ad7a56635fff\": rpc error: code = NotFound desc = could not find container \"9bbc3bc9d108614a1a523cd73a2af6df5f923fc02e52e734dc80ad7a56635fff\": container with ID starting with 9bbc3bc9d108614a1a523cd73a2af6df5f923fc02e52e734dc80ad7a56635fff not found: ID does not exist" Oct 14 15:28:29 crc kubenswrapper[4945]: I1014 15:28:29.371823 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02f9c79c-b544-43a4-a794-21293b2de0f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02f9c79c-b544-43a4-a794-21293b2de0f9" (UID: "02f9c79c-b544-43a4-a794-21293b2de0f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:28:29 crc kubenswrapper[4945]: I1014 15:28:29.465515 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f9c79c-b544-43a4-a794-21293b2de0f9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:29 crc kubenswrapper[4945]: I1014 15:28:29.556015 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vchw9"] Oct 14 15:28:29 crc kubenswrapper[4945]: I1014 15:28:29.560946 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vchw9"] Oct 14 15:28:29 crc kubenswrapper[4945]: I1014 15:28:29.932582 4945 generic.go:334] "Generic (PLEG): container finished" podID="7a990723-8b29-4962-8861-f692e2da3bcd" containerID="d775cbfc900ad9db579052a46a21f1849df1313e45345912b96b26c4468fe575" exitCode=0 Oct 14 15:28:29 crc kubenswrapper[4945]: I1014 15:28:29.932628 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpc78" event={"ID":"7a990723-8b29-4962-8861-f692e2da3bcd","Type":"ContainerDied","Data":"d775cbfc900ad9db579052a46a21f1849df1313e45345912b96b26c4468fe575"} Oct 14 15:28:30 crc kubenswrapper[4945]: I1014 15:28:30.772543 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02f9c79c-b544-43a4-a794-21293b2de0f9" path="/var/lib/kubelet/pods/02f9c79c-b544-43a4-a794-21293b2de0f9/volumes" Oct 14 15:28:31 crc kubenswrapper[4945]: I1014 15:28:31.973041 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpc78" event={"ID":"7a990723-8b29-4962-8861-f692e2da3bcd","Type":"ContainerStarted","Data":"e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511"} Oct 14 15:28:31 crc kubenswrapper[4945]: I1014 15:28:31.977089 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6" event={"ID":"795c3628-2ec0-4846-b345-3019636282e9","Type":"ContainerStarted","Data":"46f0eb7ef2c43c7b8a7ddce75215477fbaf11599ab336732bbe129702c71c23c"} Oct 14 15:28:31 crc kubenswrapper[4945]: I1014 15:28:31.977979 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6" Oct 14 15:28:31 crc kubenswrapper[4945]: I1014 15:28:31.994515 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mpc78" podStartSLOduration=2.610258993 podStartE2EDuration="8.994498497s" podCreationTimestamp="2025-10-14 15:28:23 +0000 UTC" firstStartedPulling="2025-10-14 15:28:24.887277019 +0000 UTC m=+834.871325387" lastFinishedPulling="2025-10-14 15:28:31.271516523 +0000 UTC m=+841.255564891" observedRunningTime="2025-10-14 15:28:31.992933311 +0000 UTC m=+841.976981689" watchObservedRunningTime="2025-10-14 15:28:31.994498497 +0000 UTC m=+841.978546865" Oct 14 15:28:32 crc kubenswrapper[4945]: I1014 15:28:32.030389 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6" podStartSLOduration=1.652811207 podStartE2EDuration="9.030362476s" podCreationTimestamp="2025-10-14 15:28:23 +0000 UTC" firstStartedPulling="2025-10-14 15:28:23.894468819 +0000 UTC m=+833.878517177" lastFinishedPulling="2025-10-14 15:28:31.272020078 +0000 UTC m=+841.256068446" observedRunningTime="2025-10-14 15:28:32.023505504 +0000 UTC m=+842.007553872" watchObservedRunningTime="2025-10-14 15:28:32.030362476 +0000 UTC m=+842.014410854" Oct 14 15:28:33 crc kubenswrapper[4945]: I1014 15:28:33.577948 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-674f597f57-cqpn6" Oct 14 15:28:34 crc kubenswrapper[4945]: I1014 15:28:34.118437 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:34 crc kubenswrapper[4945]: I1014 15:28:34.118833 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:34 crc kubenswrapper[4945]: I1014 15:28:34.217297 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:36 crc kubenswrapper[4945]: I1014 15:28:36.044172 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:37 crc kubenswrapper[4945]: I1014 15:28:37.585828 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpc78"] Oct 14 15:28:38 crc kubenswrapper[4945]: I1014 15:28:38.010688 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mpc78" podUID="7a990723-8b29-4962-8861-f692e2da3bcd" containerName="registry-server" containerID="cri-o://e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511" gracePeriod=2 Oct 14 15:28:38 crc kubenswrapper[4945]: I1014 15:28:38.945259 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.004999 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjhdh\" (UniqueName: \"kubernetes.io/projected/7a990723-8b29-4962-8861-f692e2da3bcd-kube-api-access-vjhdh\") pod \"7a990723-8b29-4962-8861-f692e2da3bcd\" (UID: \"7a990723-8b29-4962-8861-f692e2da3bcd\") " Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.005113 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a990723-8b29-4962-8861-f692e2da3bcd-utilities\") pod \"7a990723-8b29-4962-8861-f692e2da3bcd\" (UID: \"7a990723-8b29-4962-8861-f692e2da3bcd\") " Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.005184 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a990723-8b29-4962-8861-f692e2da3bcd-catalog-content\") pod \"7a990723-8b29-4962-8861-f692e2da3bcd\" (UID: \"7a990723-8b29-4962-8861-f692e2da3bcd\") " Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.005781 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a990723-8b29-4962-8861-f692e2da3bcd-utilities" (OuterVolumeSpecName: "utilities") pod "7a990723-8b29-4962-8861-f692e2da3bcd" (UID: "7a990723-8b29-4962-8861-f692e2da3bcd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.017107 4945 generic.go:334] "Generic (PLEG): container finished" podID="7a990723-8b29-4962-8861-f692e2da3bcd" containerID="e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511" exitCode=0 Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.017147 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpc78" event={"ID":"7a990723-8b29-4962-8861-f692e2da3bcd","Type":"ContainerDied","Data":"e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511"} Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.017171 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpc78" event={"ID":"7a990723-8b29-4962-8861-f692e2da3bcd","Type":"ContainerDied","Data":"326cd262292a88e8830c63be0dcee8cedcdf56917bb695ac46dadbed41c0197d"} Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.017188 4945 scope.go:117] "RemoveContainer" containerID="e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.017323 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpc78" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.021368 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a990723-8b29-4962-8861-f692e2da3bcd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a990723-8b29-4962-8861-f692e2da3bcd" (UID: "7a990723-8b29-4962-8861-f692e2da3bcd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.031049 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a990723-8b29-4962-8861-f692e2da3bcd-kube-api-access-vjhdh" (OuterVolumeSpecName: "kube-api-access-vjhdh") pod "7a990723-8b29-4962-8861-f692e2da3bcd" (UID: "7a990723-8b29-4962-8861-f692e2da3bcd"). InnerVolumeSpecName "kube-api-access-vjhdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.068898 4945 scope.go:117] "RemoveContainer" containerID="d775cbfc900ad9db579052a46a21f1849df1313e45345912b96b26c4468fe575" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.103866 4945 scope.go:117] "RemoveContainer" containerID="a192378b08532bfd30ad108bd03a8303e0c3dbec56697039c4633c51510a3dc9" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.106172 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjhdh\" (UniqueName: \"kubernetes.io/projected/7a990723-8b29-4962-8861-f692e2da3bcd-kube-api-access-vjhdh\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.106194 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a990723-8b29-4962-8861-f692e2da3bcd-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.106203 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a990723-8b29-4962-8861-f692e2da3bcd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.142772 4945 scope.go:117] "RemoveContainer" containerID="e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511" Oct 14 15:28:39 crc kubenswrapper[4945]: E1014 15:28:39.144964 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511\": container with ID starting with e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511 not found: ID does not exist" containerID="e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.145000 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511"} err="failed to get container status \"e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511\": rpc error: code = NotFound desc = could not find container \"e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511\": container with ID starting with e0d06253325011d46630f575b723dbaf2b4e2e53656e52c25fe7366f067d7511 not found: ID does not exist" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.145021 4945 scope.go:117] "RemoveContainer" containerID="d775cbfc900ad9db579052a46a21f1849df1313e45345912b96b26c4468fe575" Oct 14 15:28:39 crc kubenswrapper[4945]: E1014 15:28:39.148909 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d775cbfc900ad9db579052a46a21f1849df1313e45345912b96b26c4468fe575\": container with ID starting with d775cbfc900ad9db579052a46a21f1849df1313e45345912b96b26c4468fe575 not found: ID does not exist" containerID="d775cbfc900ad9db579052a46a21f1849df1313e45345912b96b26c4468fe575" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.148940 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d775cbfc900ad9db579052a46a21f1849df1313e45345912b96b26c4468fe575"} err="failed to get container status \"d775cbfc900ad9db579052a46a21f1849df1313e45345912b96b26c4468fe575\": rpc error: code = NotFound desc = could not find container \"d775cbfc900ad9db579052a46a21f1849df1313e45345912b96b26c4468fe575\": container with ID starting with d775cbfc900ad9db579052a46a21f1849df1313e45345912b96b26c4468fe575 not found: ID does not exist" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.148959 4945 scope.go:117] "RemoveContainer" containerID="a192378b08532bfd30ad108bd03a8303e0c3dbec56697039c4633c51510a3dc9" Oct 14 15:28:39 crc kubenswrapper[4945]: E1014 15:28:39.149187 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a192378b08532bfd30ad108bd03a8303e0c3dbec56697039c4633c51510a3dc9\": container with ID starting with a192378b08532bfd30ad108bd03a8303e0c3dbec56697039c4633c51510a3dc9 not found: ID does not exist" containerID="a192378b08532bfd30ad108bd03a8303e0c3dbec56697039c4633c51510a3dc9" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.149208 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a192378b08532bfd30ad108bd03a8303e0c3dbec56697039c4633c51510a3dc9"} err="failed to get container status \"a192378b08532bfd30ad108bd03a8303e0c3dbec56697039c4633c51510a3dc9\": rpc error: code = NotFound desc = could not find container \"a192378b08532bfd30ad108bd03a8303e0c3dbec56697039c4633c51510a3dc9\": container with ID starting with a192378b08532bfd30ad108bd03a8303e0c3dbec56697039c4633c51510a3dc9 not found: ID does not exist" Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.346650 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpc78"] Oct 14 15:28:39 crc kubenswrapper[4945]: I1014 15:28:39.356348 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpc78"] Oct 14 15:28:40 crc kubenswrapper[4945]: I1014 15:28:40.771906 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a990723-8b29-4962-8861-f692e2da3bcd" path="/var/lib/kubelet/pods/7a990723-8b29-4962-8861-f692e2da3bcd/volumes" Oct 14 15:28:46 crc kubenswrapper[4945]: I1014 15:28:46.795755 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:28:46 crc kubenswrapper[4945]: I1014 15:28:46.797397 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:28:46 crc kubenswrapper[4945]: I1014 15:28:46.797553 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:28:46 crc kubenswrapper[4945]: I1014 15:28:46.798321 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d2f9a019d15034c004b5841d70bbbf8261549fb60b58e5a8e606324eb14a78f"} pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 15:28:46 crc kubenswrapper[4945]: I1014 15:28:46.798465 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" containerID="cri-o://1d2f9a019d15034c004b5841d70bbbf8261549fb60b58e5a8e606324eb14a78f" gracePeriod=600 Oct 14 15:28:48 crc kubenswrapper[4945]: I1014 15:28:48.066958 4945 generic.go:334] "Generic (PLEG): container finished" podID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerID="1d2f9a019d15034c004b5841d70bbbf8261549fb60b58e5a8e606324eb14a78f" exitCode=0 Oct 14 15:28:48 crc kubenswrapper[4945]: I1014 15:28:48.067072 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerDied","Data":"1d2f9a019d15034c004b5841d70bbbf8261549fb60b58e5a8e606324eb14a78f"} Oct 14 15:28:48 crc kubenswrapper[4945]: I1014 15:28:48.067406 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerStarted","Data":"cb33e17badcd091cccc29c2b02c75756661752ad2d76d11d53ff1811c318baac"} Oct 14 15:28:48 crc kubenswrapper[4945]: I1014 15:28:48.067430 4945 scope.go:117] "RemoveContainer" containerID="8f32c2d032f723cbbb763494c9ff11fd408a6db0a40ec4819ef209acfe780ed8" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.873597 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6"] Oct 14 15:28:49 crc kubenswrapper[4945]: E1014 15:28:49.874202 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a990723-8b29-4962-8861-f692e2da3bcd" containerName="extract-content" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.874221 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a990723-8b29-4962-8861-f692e2da3bcd" containerName="extract-content" Oct 14 15:28:49 crc kubenswrapper[4945]: E1014 15:28:49.874240 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a990723-8b29-4962-8861-f692e2da3bcd" containerName="registry-server" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.874248 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a990723-8b29-4962-8861-f692e2da3bcd" containerName="registry-server" Oct 14 15:28:49 crc kubenswrapper[4945]: E1014 15:28:49.874260 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f9c79c-b544-43a4-a794-21293b2de0f9" containerName="extract-utilities" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.874269 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f9c79c-b544-43a4-a794-21293b2de0f9" containerName="extract-utilities" Oct 14 15:28:49 crc kubenswrapper[4945]: E1014 15:28:49.874280 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f9c79c-b544-43a4-a794-21293b2de0f9" containerName="extract-content" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.874286 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f9c79c-b544-43a4-a794-21293b2de0f9" containerName="extract-content" Oct 14 15:28:49 crc kubenswrapper[4945]: E1014 15:28:49.874299 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f9c79c-b544-43a4-a794-21293b2de0f9" containerName="registry-server" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.874306 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f9c79c-b544-43a4-a794-21293b2de0f9" containerName="registry-server" Oct 14 15:28:49 crc kubenswrapper[4945]: E1014 15:28:49.874323 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a990723-8b29-4962-8861-f692e2da3bcd" containerName="extract-utilities" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.874329 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a990723-8b29-4962-8861-f692e2da3bcd" containerName="extract-utilities" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.874478 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a990723-8b29-4962-8861-f692e2da3bcd" containerName="registry-server" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.874496 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="02f9c79c-b544-43a4-a794-21293b2de0f9" containerName="registry-server" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.875336 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.880940 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-qqj8g" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.885585 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6"] Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.897296 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j"] Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.898520 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.899714 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t"] Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.900337 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-vg7lm" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.900737 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.909054 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-smqf9" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.922709 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j"] Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.933586 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t"] Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.949392 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x"] Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.950530 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.953279 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr"] Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.954359 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.962283 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-t4lhb" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.962546 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-btj2k" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.963102 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4wv8\" (UniqueName: \"kubernetes.io/projected/86d698df-b089-498a-98e5-65c196aabcc3-kube-api-access-v4wv8\") pod \"glance-operator-controller-manager-7bb46cd7d-6fg2j\" (UID: \"86d698df-b089-498a-98e5-65c196aabcc3\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.963158 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvh5b\" (UniqueName: \"kubernetes.io/projected/f5ac9c23-4d14-460d-9255-16d1865e9aa1-kube-api-access-wvh5b\") pod \"barbican-operator-controller-manager-64f84fcdbb-xlwn6\" (UID: \"f5ac9c23-4d14-460d-9255-16d1865e9aa1\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.963202 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b7hn\" (UniqueName: \"kubernetes.io/projected/64596f58-bc5a-4cdf-80d4-22002c0c9575-kube-api-access-9b7hn\") pod \"designate-operator-controller-manager-687df44cdb-gf66t\" (UID: \"64596f58-bc5a-4cdf-80d4-22002c0c9575\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t" Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.966969 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x"] Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.979635 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt"] Oct 14 15:28:49 crc kubenswrapper[4945]: I1014 15:28:49.980628 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.009825 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-l6v8t" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.061355 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.064640 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4wv8\" (UniqueName: \"kubernetes.io/projected/86d698df-b089-498a-98e5-65c196aabcc3-kube-api-access-v4wv8\") pod \"glance-operator-controller-manager-7bb46cd7d-6fg2j\" (UID: \"86d698df-b089-498a-98e5-65c196aabcc3\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.064693 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvh5b\" (UniqueName: \"kubernetes.io/projected/f5ac9c23-4d14-460d-9255-16d1865e9aa1-kube-api-access-wvh5b\") pod \"barbican-operator-controller-manager-64f84fcdbb-xlwn6\" (UID: \"f5ac9c23-4d14-460d-9255-16d1865e9aa1\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.064728 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b7hn\" (UniqueName: \"kubernetes.io/projected/64596f58-bc5a-4cdf-80d4-22002c0c9575-kube-api-access-9b7hn\") pod \"designate-operator-controller-manager-687df44cdb-gf66t\" (UID: \"64596f58-bc5a-4cdf-80d4-22002c0c9575\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.064755 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmqbl\" (UniqueName: \"kubernetes.io/projected/0a17b021-e8aa-4f88-b150-596885e48ab1-kube-api-access-pmqbl\") pod \"heat-operator-controller-manager-6d9967f8dd-v9hhr\" (UID: \"0a17b021-e8aa-4f88-b150-596885e48ab1\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.064777 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4hdg\" (UniqueName: \"kubernetes.io/projected/80c510fc-48a8-44c1-9927-c9b0ce934410-kube-api-access-l4hdg\") pod \"horizon-operator-controller-manager-6d74794d9b-wchqt\" (UID: \"80c510fc-48a8-44c1-9927-c9b0ce934410\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.064830 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swslg\" (UniqueName: \"kubernetes.io/projected/3ae6c6f4-2d25-420b-8149-a976de545cd6-kube-api-access-swslg\") pod \"cinder-operator-controller-manager-59cdc64769-8gm4x\" (UID: \"3ae6c6f4-2d25-420b-8149-a976de545cd6\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.087969 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.100683 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.101279 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4wv8\" (UniqueName: \"kubernetes.io/projected/86d698df-b089-498a-98e5-65c196aabcc3-kube-api-access-v4wv8\") pod \"glance-operator-controller-manager-7bb46cd7d-6fg2j\" (UID: \"86d698df-b089-498a-98e5-65c196aabcc3\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.102677 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.109479 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b7hn\" (UniqueName: \"kubernetes.io/projected/64596f58-bc5a-4cdf-80d4-22002c0c9575-kube-api-access-9b7hn\") pod \"designate-operator-controller-manager-687df44cdb-gf66t\" (UID: \"64596f58-bc5a-4cdf-80d4-22002c0c9575\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.111255 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.112377 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.116004 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.117306 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.123704 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvh5b\" (UniqueName: \"kubernetes.io/projected/f5ac9c23-4d14-460d-9255-16d1865e9aa1-kube-api-access-wvh5b\") pod \"barbican-operator-controller-manager-64f84fcdbb-xlwn6\" (UID: \"f5ac9c23-4d14-460d-9255-16d1865e9aa1\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.128114 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.128346 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-l65lr" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.128490 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-hzc5k" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.128930 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.129044 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-sks7j" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.149009 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.159686 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.165728 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.166568 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmqbl\" (UniqueName: \"kubernetes.io/projected/0a17b021-e8aa-4f88-b150-596885e48ab1-kube-api-access-pmqbl\") pod \"heat-operator-controller-manager-6d9967f8dd-v9hhr\" (UID: \"0a17b021-e8aa-4f88-b150-596885e48ab1\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.166664 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4hdg\" (UniqueName: \"kubernetes.io/projected/80c510fc-48a8-44c1-9927-c9b0ce934410-kube-api-access-l4hdg\") pod \"horizon-operator-controller-manager-6d74794d9b-wchqt\" (UID: \"80c510fc-48a8-44c1-9927-c9b0ce934410\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.166786 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk8h7\" (UniqueName: \"kubernetes.io/projected/20fbc887-9299-460d-9890-f9d7e0235118-kube-api-access-pk8h7\") pod \"ironic-operator-controller-manager-76ccf76b65-p9trl\" (UID: \"20fbc887-9299-460d-9890-f9d7e0235118\") " pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.166910 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41d187ec-0816-4800-9d76-9d7a3f364391-cert\") pod \"infra-operator-controller-manager-585fc5b659-pk25j\" (UID: \"41d187ec-0816-4800-9d76-9d7a3f364391\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.167010 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dhmw\" (UniqueName: \"kubernetes.io/projected/97138e7c-7207-4c47-898f-f71625e0e49b-kube-api-access-6dhmw\") pod \"keystone-operator-controller-manager-ddb98f99b-fqb29\" (UID: \"97138e7c-7207-4c47-898f-f71625e0e49b\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.167109 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np7lh\" (UniqueName: \"kubernetes.io/projected/41d187ec-0816-4800-9d76-9d7a3f364391-kube-api-access-np7lh\") pod \"infra-operator-controller-manager-585fc5b659-pk25j\" (UID: \"41d187ec-0816-4800-9d76-9d7a3f364391\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.167197 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swslg\" (UniqueName: \"kubernetes.io/projected/3ae6c6f4-2d25-420b-8149-a976de545cd6-kube-api-access-swslg\") pod \"cinder-operator-controller-manager-59cdc64769-8gm4x\" (UID: \"3ae6c6f4-2d25-420b-8149-a976de545cd6\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.166971 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.172190 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-9knb7" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.181703 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.183050 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.186513 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-46svt" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.189680 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.190858 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.193140 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-w7hs4" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.193205 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.200095 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.220758 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.221860 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4hdg\" (UniqueName: \"kubernetes.io/projected/80c510fc-48a8-44c1-9927-c9b0ce934410-kube-api-access-l4hdg\") pod \"horizon-operator-controller-manager-6d74794d9b-wchqt\" (UID: \"80c510fc-48a8-44c1-9927-c9b0ce934410\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.221910 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swslg\" (UniqueName: \"kubernetes.io/projected/3ae6c6f4-2d25-420b-8149-a976de545cd6-kube-api-access-swslg\") pod \"cinder-operator-controller-manager-59cdc64769-8gm4x\" (UID: \"3ae6c6f4-2d25-420b-8149-a976de545cd6\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.222092 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmqbl\" (UniqueName: \"kubernetes.io/projected/0a17b021-e8aa-4f88-b150-596885e48ab1-kube-api-access-pmqbl\") pod \"heat-operator-controller-manager-6d9967f8dd-v9hhr\" (UID: \"0a17b021-e8aa-4f88-b150-596885e48ab1\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.222687 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.230955 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.234479 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.238408 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.239246 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.243942 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.248929 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.250393 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.251421 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-fpl9t" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.252154 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-lfqgf" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.270454 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5f8s\" (UniqueName: \"kubernetes.io/projected/b4e68062-2195-43f9-a6ae-a58d12bd7ce2-kube-api-access-k5f8s\") pod \"nova-operator-controller-manager-57bb74c7bf-fl79b\" (UID: \"b4e68062-2195-43f9-a6ae-a58d12bd7ce2\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.270527 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dbdf\" (UniqueName: \"kubernetes.io/projected/d39181a7-852e-402c-9912-7c81a0952e90-kube-api-access-7dbdf\") pod \"mariadb-operator-controller-manager-5777b4f897-h2p64\" (UID: \"d39181a7-852e-402c-9912-7c81a0952e90\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.270570 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk8h7\" (UniqueName: \"kubernetes.io/projected/20fbc887-9299-460d-9890-f9d7e0235118-kube-api-access-pk8h7\") pod \"ironic-operator-controller-manager-76ccf76b65-p9trl\" (UID: \"20fbc887-9299-460d-9890-f9d7e0235118\") " pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.270602 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c26nq\" (UniqueName: \"kubernetes.io/projected/d10e063a-f99e-4fc8-b539-224bbf23f466-kube-api-access-c26nq\") pod \"neutron-operator-controller-manager-797d478b46-qx6hl\" (UID: \"d10e063a-f99e-4fc8-b539-224bbf23f466\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.270620 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41d187ec-0816-4800-9d76-9d7a3f364391-cert\") pod \"infra-operator-controller-manager-585fc5b659-pk25j\" (UID: \"41d187ec-0816-4800-9d76-9d7a3f364391\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.270650 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dhmw\" (UniqueName: \"kubernetes.io/projected/97138e7c-7207-4c47-898f-f71625e0e49b-kube-api-access-6dhmw\") pod \"keystone-operator-controller-manager-ddb98f99b-fqb29\" (UID: \"97138e7c-7207-4c47-898f-f71625e0e49b\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.270667 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgwnx\" (UniqueName: \"kubernetes.io/projected/2a652464-b536-4402-a883-f52390e0327b-kube-api-access-qgwnx\") pod \"manila-operator-controller-manager-59578bc799-jcrs7\" (UID: \"2a652464-b536-4402-a883-f52390e0327b\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.270689 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np7lh\" (UniqueName: \"kubernetes.io/projected/41d187ec-0816-4800-9d76-9d7a3f364391-kube-api-access-np7lh\") pod \"infra-operator-controller-manager-585fc5b659-pk25j\" (UID: \"41d187ec-0816-4800-9d76-9d7a3f364391\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" Oct 14 15:28:50 crc kubenswrapper[4945]: E1014 15:28:50.271141 4945 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 14 15:28:50 crc kubenswrapper[4945]: E1014 15:28:50.271184 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41d187ec-0816-4800-9d76-9d7a3f364391-cert podName:41d187ec-0816-4800-9d76-9d7a3f364391 nodeName:}" failed. No retries permitted until 2025-10-14 15:28:50.771169717 +0000 UTC m=+860.755218085 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/41d187ec-0816-4800-9d76-9d7a3f364391-cert") pod "infra-operator-controller-manager-585fc5b659-pk25j" (UID: "41d187ec-0816-4800-9d76-9d7a3f364391") : secret "infra-operator-webhook-server-cert" not found Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.275902 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.276901 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.282690 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.283694 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.286297 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.291692 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-tdskf" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.294828 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np7lh\" (UniqueName: \"kubernetes.io/projected/41d187ec-0816-4800-9d76-9d7a3f364391-kube-api-access-np7lh\") pod \"infra-operator-controller-manager-585fc5b659-pk25j\" (UID: \"41d187ec-0816-4800-9d76-9d7a3f364391\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.295148 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk8h7\" (UniqueName: \"kubernetes.io/projected/20fbc887-9299-460d-9890-f9d7e0235118-kube-api-access-pk8h7\") pod \"ironic-operator-controller-manager-76ccf76b65-p9trl\" (UID: \"20fbc887-9299-460d-9890-f9d7e0235118\") " pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.298502 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.299442 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.304429 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dhmw\" (UniqueName: \"kubernetes.io/projected/97138e7c-7207-4c47-898f-f71625e0e49b-kube-api-access-6dhmw\") pod \"keystone-operator-controller-manager-ddb98f99b-fqb29\" (UID: \"97138e7c-7207-4c47-898f-f71625e0e49b\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.305259 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-8vlkr" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.308552 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-58vsj"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.309687 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.310799 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.314178 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-chn7h" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.316606 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.324722 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.328331 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.329495 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.332813 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-m7nvv" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.335833 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-58vsj"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.350120 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.357492 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.371743 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5f8s\" (UniqueName: \"kubernetes.io/projected/b4e68062-2195-43f9-a6ae-a58d12bd7ce2-kube-api-access-k5f8s\") pod \"nova-operator-controller-manager-57bb74c7bf-fl79b\" (UID: \"b4e68062-2195-43f9-a6ae-a58d12bd7ce2\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.371788 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jl6p\" (UniqueName: \"kubernetes.io/projected/4af90a70-780d-4d6f-9fc5-6a384002a606-kube-api-access-8jl6p\") pod \"octavia-operator-controller-manager-6d7c7ddf95-vspm4\" (UID: \"4af90a70-780d-4d6f-9fc5-6a384002a606\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.371832 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnckw\" (UniqueName: \"kubernetes.io/projected/37b893c7-b6a7-4a19-bc3e-b251deacdd27-kube-api-access-xnckw\") pod \"ovn-operator-controller-manager-869cc7797f-r6tvw\" (UID: \"37b893c7-b6a7-4a19-bc3e-b251deacdd27\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.371847 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlvxs\" (UniqueName: \"kubernetes.io/projected/e3193e9c-898b-456d-a825-ea8176568803-kube-api-access-zlvxs\") pod \"swift-operator-controller-manager-5f4d5dfdc6-48x4h\" (UID: \"e3193e9c-898b-456d-a825-ea8176568803\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.371902 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dbdf\" (UniqueName: \"kubernetes.io/projected/d39181a7-852e-402c-9912-7c81a0952e90-kube-api-access-7dbdf\") pod \"mariadb-operator-controller-manager-5777b4f897-h2p64\" (UID: \"d39181a7-852e-402c-9912-7c81a0952e90\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.371927 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnwcd\" (UniqueName: \"kubernetes.io/projected/669db3c1-3b7e-4bbd-980e-93fdea61a1cc-kube-api-access-mnwcd\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w\" (UID: \"669db3c1-3b7e-4bbd-980e-93fdea61a1cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.371970 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/669db3c1-3b7e-4bbd-980e-93fdea61a1cc-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w\" (UID: \"669db3c1-3b7e-4bbd-980e-93fdea61a1cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.371990 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c26nq\" (UniqueName: \"kubernetes.io/projected/d10e063a-f99e-4fc8-b539-224bbf23f466-kube-api-access-c26nq\") pod \"neutron-operator-controller-manager-797d478b46-qx6hl\" (UID: \"d10e063a-f99e-4fc8-b539-224bbf23f466\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.372021 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgwnx\" (UniqueName: \"kubernetes.io/projected/2a652464-b536-4402-a883-f52390e0327b-kube-api-access-qgwnx\") pod \"manila-operator-controller-manager-59578bc799-jcrs7\" (UID: \"2a652464-b536-4402-a883-f52390e0327b\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.372042 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpzz9\" (UniqueName: \"kubernetes.io/projected/01b34d70-1617-4e4d-81ab-8cada4bb1edf-kube-api-access-dpzz9\") pod \"placement-operator-controller-manager-664664cb68-58vsj\" (UID: \"01b34d70-1617-4e4d-81ab-8cada4bb1edf\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.393726 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c26nq\" (UniqueName: \"kubernetes.io/projected/d10e063a-f99e-4fc8-b539-224bbf23f466-kube-api-access-c26nq\") pod \"neutron-operator-controller-manager-797d478b46-qx6hl\" (UID: \"d10e063a-f99e-4fc8-b539-224bbf23f466\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.394568 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5f8s\" (UniqueName: \"kubernetes.io/projected/b4e68062-2195-43f9-a6ae-a58d12bd7ce2-kube-api-access-k5f8s\") pod \"nova-operator-controller-manager-57bb74c7bf-fl79b\" (UID: \"b4e68062-2195-43f9-a6ae-a58d12bd7ce2\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.394698 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dbdf\" (UniqueName: \"kubernetes.io/projected/d39181a7-852e-402c-9912-7c81a0952e90-kube-api-access-7dbdf\") pod \"mariadb-operator-controller-manager-5777b4f897-h2p64\" (UID: \"d39181a7-852e-402c-9912-7c81a0952e90\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.401618 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgwnx\" (UniqueName: \"kubernetes.io/projected/2a652464-b536-4402-a883-f52390e0327b-kube-api-access-qgwnx\") pod \"manila-operator-controller-manager-59578bc799-jcrs7\" (UID: \"2a652464-b536-4402-a883-f52390e0327b\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.403366 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.433798 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-48769"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.438853 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.441335 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-fv4pm" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.486333 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.488148 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jl6p\" (UniqueName: \"kubernetes.io/projected/4af90a70-780d-4d6f-9fc5-6a384002a606-kube-api-access-8jl6p\") pod \"octavia-operator-controller-manager-6d7c7ddf95-vspm4\" (UID: \"4af90a70-780d-4d6f-9fc5-6a384002a606\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.488224 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnckw\" (UniqueName: \"kubernetes.io/projected/37b893c7-b6a7-4a19-bc3e-b251deacdd27-kube-api-access-xnckw\") pod \"ovn-operator-controller-manager-869cc7797f-r6tvw\" (UID: \"37b893c7-b6a7-4a19-bc3e-b251deacdd27\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.488248 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlvxs\" (UniqueName: \"kubernetes.io/projected/e3193e9c-898b-456d-a825-ea8176568803-kube-api-access-zlvxs\") pod \"swift-operator-controller-manager-5f4d5dfdc6-48x4h\" (UID: \"e3193e9c-898b-456d-a825-ea8176568803\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.488285 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnwcd\" (UniqueName: \"kubernetes.io/projected/669db3c1-3b7e-4bbd-980e-93fdea61a1cc-kube-api-access-mnwcd\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w\" (UID: \"669db3c1-3b7e-4bbd-980e-93fdea61a1cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.488329 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdmsc\" (UniqueName: \"kubernetes.io/projected/50453e84-c640-4a80-ac7d-0fe2ca24acfd-kube-api-access-sdmsc\") pod \"telemetry-operator-controller-manager-578874c84d-48769\" (UID: \"50453e84-c640-4a80-ac7d-0fe2ca24acfd\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.488365 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/669db3c1-3b7e-4bbd-980e-93fdea61a1cc-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w\" (UID: \"669db3c1-3b7e-4bbd-980e-93fdea61a1cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.488430 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpzz9\" (UniqueName: \"kubernetes.io/projected/01b34d70-1617-4e4d-81ab-8cada4bb1edf-kube-api-access-dpzz9\") pod \"placement-operator-controller-manager-664664cb68-58vsj\" (UID: \"01b34d70-1617-4e4d-81ab-8cada4bb1edf\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" Oct 14 15:28:50 crc kubenswrapper[4945]: E1014 15:28:50.489836 4945 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 14 15:28:50 crc kubenswrapper[4945]: E1014 15:28:50.491990 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/669db3c1-3b7e-4bbd-980e-93fdea61a1cc-cert podName:669db3c1-3b7e-4bbd-980e-93fdea61a1cc nodeName:}" failed. No retries permitted until 2025-10-14 15:28:50.989891784 +0000 UTC m=+860.973940142 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/669db3c1-3b7e-4bbd-980e-93fdea61a1cc-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" (UID: "669db3c1-3b7e-4bbd-980e-93fdea61a1cc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.493392 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-48769"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.516115 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpzz9\" (UniqueName: \"kubernetes.io/projected/01b34d70-1617-4e4d-81ab-8cada4bb1edf-kube-api-access-dpzz9\") pod \"placement-operator-controller-manager-664664cb68-58vsj\" (UID: \"01b34d70-1617-4e4d-81ab-8cada4bb1edf\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.522897 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.545774 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnckw\" (UniqueName: \"kubernetes.io/projected/37b893c7-b6a7-4a19-bc3e-b251deacdd27-kube-api-access-xnckw\") pod \"ovn-operator-controller-manager-869cc7797f-r6tvw\" (UID: \"37b893c7-b6a7-4a19-bc3e-b251deacdd27\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.550424 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnwcd\" (UniqueName: \"kubernetes.io/projected/669db3c1-3b7e-4bbd-980e-93fdea61a1cc-kube-api-access-mnwcd\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w\" (UID: \"669db3c1-3b7e-4bbd-980e-93fdea61a1cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.551335 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlvxs\" (UniqueName: \"kubernetes.io/projected/e3193e9c-898b-456d-a825-ea8176568803-kube-api-access-zlvxs\") pod \"swift-operator-controller-manager-5f4d5dfdc6-48x4h\" (UID: \"e3193e9c-898b-456d-a825-ea8176568803\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.552499 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jl6p\" (UniqueName: \"kubernetes.io/projected/4af90a70-780d-4d6f-9fc5-6a384002a606-kube-api-access-8jl6p\") pod \"octavia-operator-controller-manager-6d7c7ddf95-vspm4\" (UID: \"4af90a70-780d-4d6f-9fc5-6a384002a606\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.580815 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.582692 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.590023 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-bftbq" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.591450 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.592260 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdbgd\" (UniqueName: \"kubernetes.io/projected/2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c-kube-api-access-hdbgd\") pod \"test-operator-controller-manager-ffcdd6c94-kkqht\" (UID: \"2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.592350 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdmsc\" (UniqueName: \"kubernetes.io/projected/50453e84-c640-4a80-ac7d-0fe2ca24acfd-kube-api-access-sdmsc\") pod \"telemetry-operator-controller-manager-578874c84d-48769\" (UID: \"50453e84-c640-4a80-ac7d-0fe2ca24acfd\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.596408 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.618832 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.620138 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.630455 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.631335 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdmsc\" (UniqueName: \"kubernetes.io/projected/50453e84-c640-4a80-ac7d-0fe2ca24acfd-kube-api-access-sdmsc\") pod \"telemetry-operator-controller-manager-578874c84d-48769\" (UID: \"50453e84-c640-4a80-ac7d-0fe2ca24acfd\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.638968 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.640805 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.646367 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-59bfx" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.646787 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.657365 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.673134 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.686234 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.693500 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpbtq\" (UniqueName: \"kubernetes.io/projected/2c954bb2-bf12-491f-92e8-637e0eb13ad5-kube-api-access-dpbtq\") pod \"watcher-operator-controller-manager-646675d848-sjvtx\" (UID: \"2c954bb2-bf12-491f-92e8-637e0eb13ad5\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.693557 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdbgd\" (UniqueName: \"kubernetes.io/projected/2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c-kube-api-access-hdbgd\") pod \"test-operator-controller-manager-ffcdd6c94-kkqht\" (UID: \"2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.709822 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.712854 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.717064 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.717936 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-q9dpq" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.722357 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdbgd\" (UniqueName: \"kubernetes.io/projected/2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c-kube-api-access-hdbgd\") pod \"test-operator-controller-manager-ffcdd6c94-kkqht\" (UID: \"2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.722645 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.725146 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.768283 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.769706 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.783709 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-s48z6" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.790249 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.796475 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41d187ec-0816-4800-9d76-9d7a3f364391-cert\") pod \"infra-operator-controller-manager-585fc5b659-pk25j\" (UID: \"41d187ec-0816-4800-9d76-9d7a3f364391\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.796601 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpbtq\" (UniqueName: \"kubernetes.io/projected/2c954bb2-bf12-491f-92e8-637e0eb13ad5-kube-api-access-dpbtq\") pod \"watcher-operator-controller-manager-646675d848-sjvtx\" (UID: \"2c954bb2-bf12-491f-92e8-637e0eb13ad5\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.803665 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41d187ec-0816-4800-9d76-9d7a3f364391-cert\") pod \"infra-operator-controller-manager-585fc5b659-pk25j\" (UID: \"41d187ec-0816-4800-9d76-9d7a3f364391\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.826384 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6"] Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.839322 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpbtq\" (UniqueName: \"kubernetes.io/projected/2c954bb2-bf12-491f-92e8-637e0eb13ad5-kube-api-access-dpbtq\") pod \"watcher-operator-controller-manager-646675d848-sjvtx\" (UID: \"2c954bb2-bf12-491f-92e8-637e0eb13ad5\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.898751 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxvbx\" (UniqueName: \"kubernetes.io/projected/26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad-kube-api-access-fxvbx\") pod \"openstack-operator-controller-manager-b9d6d9595-9zj7r\" (UID: \"26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad\") " pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.900067 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gqx8\" (UniqueName: \"kubernetes.io/projected/f23ed77e-a951-4003-bbc2-a2e06431533e-kube-api-access-5gqx8\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q\" (UID: \"f23ed77e-a951-4003-bbc2-a2e06431533e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q" Oct 14 15:28:50 crc kubenswrapper[4945]: I1014 15:28:50.900123 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad-cert\") pod \"openstack-operator-controller-manager-b9d6d9595-9zj7r\" (UID: \"26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad\") " pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" Oct 14 15:28:50 crc kubenswrapper[4945]: W1014 15:28:50.922108 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5ac9c23_4d14_460d_9255_16d1865e9aa1.slice/crio-1b62aee8cf272f81a95c5bf1c5d49052a5c9344f8fa63858acea0fea2469425c WatchSource:0}: Error finding container 1b62aee8cf272f81a95c5bf1c5d49052a5c9344f8fa63858acea0fea2469425c: Status 404 returned error can't find the container with id 1b62aee8cf272f81a95c5bf1c5d49052a5c9344f8fa63858acea0fea2469425c Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.003726 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxvbx\" (UniqueName: \"kubernetes.io/projected/26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad-kube-api-access-fxvbx\") pod \"openstack-operator-controller-manager-b9d6d9595-9zj7r\" (UID: \"26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad\") " pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.003788 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gqx8\" (UniqueName: \"kubernetes.io/projected/f23ed77e-a951-4003-bbc2-a2e06431533e-kube-api-access-5gqx8\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q\" (UID: \"f23ed77e-a951-4003-bbc2-a2e06431533e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q" Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.003818 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad-cert\") pod \"openstack-operator-controller-manager-b9d6d9595-9zj7r\" (UID: \"26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad\") " pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.003894 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/669db3c1-3b7e-4bbd-980e-93fdea61a1cc-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w\" (UID: \"669db3c1-3b7e-4bbd-980e-93fdea61a1cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" Oct 14 15:28:51 crc kubenswrapper[4945]: E1014 15:28:51.004778 4945 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 14 15:28:51 crc kubenswrapper[4945]: E1014 15:28:51.004850 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad-cert podName:26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad nodeName:}" failed. No retries permitted until 2025-10-14 15:28:51.504834505 +0000 UTC m=+861.488882863 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad-cert") pod "openstack-operator-controller-manager-b9d6d9595-9zj7r" (UID: "26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad") : secret "webhook-server-cert" not found Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.022236 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.024563 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/669db3c1-3b7e-4bbd-980e-93fdea61a1cc-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w\" (UID: \"669db3c1-3b7e-4bbd-980e-93fdea61a1cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.030272 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxvbx\" (UniqueName: \"kubernetes.io/projected/26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad-kube-api-access-fxvbx\") pod \"openstack-operator-controller-manager-b9d6d9595-9zj7r\" (UID: \"26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad\") " pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.032051 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gqx8\" (UniqueName: \"kubernetes.io/projected/f23ed77e-a951-4003-bbc2-a2e06431533e-kube-api-access-5gqx8\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q\" (UID: \"f23ed77e-a951-4003-bbc2-a2e06431533e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q" Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.036852 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.059273 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.100330 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.185794 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q" Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.195792 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6" event={"ID":"f5ac9c23-4d14-460d-9255-16d1865e9aa1","Type":"ContainerStarted","Data":"1b62aee8cf272f81a95c5bf1c5d49052a5c9344f8fa63858acea0fea2469425c"} Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.291153 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t"] Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.309457 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j"] Oct 14 15:28:51 crc kubenswrapper[4945]: W1014 15:28:51.329838 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86d698df_b089_498a_98e5_65c196aabcc3.slice/crio-1b8712441f344c7af5870c1fd4d3157097a1073822570b915bceb2d4e48ad31d WatchSource:0}: Error finding container 1b8712441f344c7af5870c1fd4d3157097a1073822570b915bceb2d4e48ad31d: Status 404 returned error can't find the container with id 1b8712441f344c7af5870c1fd4d3157097a1073822570b915bceb2d4e48ad31d Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.520431 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad-cert\") pod \"openstack-operator-controller-manager-b9d6d9595-9zj7r\" (UID: \"26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad\") " pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" Oct 14 15:28:51 crc kubenswrapper[4945]: E1014 15:28:51.520923 4945 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 14 15:28:51 crc kubenswrapper[4945]: E1014 15:28:51.521012 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad-cert podName:26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad nodeName:}" failed. No retries permitted until 2025-10-14 15:28:52.520987103 +0000 UTC m=+862.505035471 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad-cert") pod "openstack-operator-controller-manager-b9d6d9595-9zj7r" (UID: "26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad") : secret "webhook-server-cert" not found Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.590372 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr"] Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.600867 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b"] Oct 14 15:28:51 crc kubenswrapper[4945]: W1014 15:28:51.621272 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ae6c6f4_2d25_420b_8149_a976de545cd6.slice/crio-8c8a8df7f45227659ffd2f675ab815258c93b808797e8730f038e4766f2e284e WatchSource:0}: Error finding container 8c8a8df7f45227659ffd2f675ab815258c93b808797e8730f038e4766f2e284e: Status 404 returned error can't find the container with id 8c8a8df7f45227659ffd2f675ab815258c93b808797e8730f038e4766f2e284e Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.624228 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl"] Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.635596 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x"] Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.643652 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt"] Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.665469 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7"] Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.669973 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw"] Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.673326 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29"] Oct 14 15:28:51 crc kubenswrapper[4945]: W1014 15:28:51.688148 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37b893c7_b6a7_4a19_bc3e_b251deacdd27.slice/crio-27a6a236d493cd3a82ee7b6725e28c4f7b559acb3f174310fab1d40b10525700 WatchSource:0}: Error finding container 27a6a236d493cd3a82ee7b6725e28c4f7b559acb3f174310fab1d40b10525700: Status 404 returned error can't find the container with id 27a6a236d493cd3a82ee7b6725e28c4f7b559acb3f174310fab1d40b10525700 Oct 14 15:28:51 crc kubenswrapper[4945]: W1014 15:28:51.691096 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97138e7c_7207_4c47_898f_f71625e0e49b.slice/crio-fe18a3548d7192a4b854c1427f05df3dfd37d36d90ac453e0c81564e560ece9d WatchSource:0}: Error finding container fe18a3548d7192a4b854c1427f05df3dfd37d36d90ac453e0c81564e560ece9d: Status 404 returned error can't find the container with id fe18a3548d7192a4b854c1427f05df3dfd37d36d90ac453e0c81564e560ece9d Oct 14 15:28:51 crc kubenswrapper[4945]: I1014 15:28:51.998044 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h"] Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.003655 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-58vsj"] Oct 14 15:28:52 crc kubenswrapper[4945]: W1014 15:28:52.021002 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01b34d70_1617_4e4d_81ab_8cada4bb1edf.slice/crio-35d7111e42d06e047a4790fd7e9c56aed9d1fc19012c7867ae1fed8da116318e WatchSource:0}: Error finding container 35d7111e42d06e047a4790fd7e9c56aed9d1fc19012c7867ae1fed8da116318e: Status 404 returned error can't find the container with id 35d7111e42d06e047a4790fd7e9c56aed9d1fc19012c7867ae1fed8da116318e Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.034076 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q"] Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.040852 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64"] Oct 14 15:28:52 crc kubenswrapper[4945]: W1014 15:28:52.041757 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf23ed77e_a951_4003_bbc2_a2e06431533e.slice/crio-7348d031f2edbb4ce2dc15b7793fa3c173bbfd7cdb5ac4aeacca99df2096f024 WatchSource:0}: Error finding container 7348d031f2edbb4ce2dc15b7793fa3c173bbfd7cdb5ac4aeacca99df2096f024: Status 404 returned error can't find the container with id 7348d031f2edbb4ce2dc15b7793fa3c173bbfd7cdb5ac4aeacca99df2096f024 Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.048636 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5gqx8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q_openstack-operators(f23ed77e-a951-4003-bbc2-a2e06431533e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.048629 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c26nq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-797d478b46-qx6hl_openstack-operators(d10e063a-f99e-4fc8-b539-224bbf23f466): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.050494 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q" podUID="f23ed77e-a951-4003-bbc2-a2e06431533e" Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.055993 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4"] Oct 14 15:28:52 crc kubenswrapper[4945]: W1014 15:28:52.057239 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c954bb2_bf12_491f_92e8_637e0eb13ad5.slice/crio-f256220827a9d244752aea8731e1e85be2b272afbce4fb627533a04e6b10dc8b WatchSource:0}: Error finding container f256220827a9d244752aea8731e1e85be2b272afbce4fb627533a04e6b10dc8b: Status 404 returned error can't find the container with id f256220827a9d244752aea8731e1e85be2b272afbce4fb627533a04e6b10dc8b Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.059280 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hdbgd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-ffcdd6c94-kkqht_openstack-operators(2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.063961 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-48769"] Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.065262 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sdmsc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-578874c84d-48769_openstack-operators(50453e84-c640-4a80-ac7d-0fe2ca24acfd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.068800 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7dbdf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-5777b4f897-h2p64_openstack-operators(d39181a7-852e-402c-9912-7c81a0952e90): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.069339 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx"] Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.073130 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dpbtq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-646675d848-sjvtx_openstack-operators(2c954bb2-bf12-491f-92e8-637e0eb13ad5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.074438 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl"] Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.091055 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht"] Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.207929 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j"] Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.223353 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w"] Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.319989 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" event={"ID":"01b34d70-1617-4e4d-81ab-8cada4bb1edf","Type":"ContainerStarted","Data":"35d7111e42d06e047a4790fd7e9c56aed9d1fc19012c7867ae1fed8da116318e"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.327706 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" event={"ID":"2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c","Type":"ContainerStarted","Data":"009b27979ddce2a6fd072f50c63766169deeb3f65630529cc4d443889f179ab9"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.335744 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" event={"ID":"20fbc887-9299-460d-9890-f9d7e0235118","Type":"ContainerStarted","Data":"97aa91ba26440fe00eeb0964d936114d99ff34faa9fce18f5349d5f28b1ae3c4"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.336942 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" event={"ID":"d39181a7-852e-402c-9912-7c81a0952e90","Type":"ContainerStarted","Data":"d715bb2548df8ab028237210e6210b3a91130e08551afe4ae3ac433f125230d1"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.338391 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t" event={"ID":"64596f58-bc5a-4cdf-80d4-22002c0c9575","Type":"ContainerStarted","Data":"dd08ae763f6e53f3f6d3c602cee620310a1e4b049bdb4480ef965cada1473bda"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.339729 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" event={"ID":"d10e063a-f99e-4fc8-b539-224bbf23f466","Type":"ContainerStarted","Data":"06b1d9343573b1ddea788b0ebff0df4086071543fa2e71622e06cab3f0767ef1"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.343796 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" event={"ID":"97138e7c-7207-4c47-898f-f71625e0e49b","Type":"ContainerStarted","Data":"fe18a3548d7192a4b854c1427f05df3dfd37d36d90ac453e0c81564e560ece9d"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.347160 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7" event={"ID":"2a652464-b536-4402-a883-f52390e0327b","Type":"ContainerStarted","Data":"055ec948313b89f61b5755114c509a033f3c34ffd56b77f89557dcba159251d1"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.364770 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" event={"ID":"b4e68062-2195-43f9-a6ae-a58d12bd7ce2","Type":"ContainerStarted","Data":"03528512f859ce63a36dd839cf4677cc1aa9d7929856037d1c33c2e2f4779b23"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.384915 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4" event={"ID":"4af90a70-780d-4d6f-9fc5-6a384002a606","Type":"ContainerStarted","Data":"a96d640075f251ff5af2f4986e0b520970fe80c520f2bda07bfd40b93c910790"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.386406 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j" event={"ID":"86d698df-b089-498a-98e5-65c196aabcc3","Type":"ContainerStarted","Data":"1b8712441f344c7af5870c1fd4d3157097a1073822570b915bceb2d4e48ad31d"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.387485 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" event={"ID":"2c954bb2-bf12-491f-92e8-637e0eb13ad5","Type":"ContainerStarted","Data":"f256220827a9d244752aea8731e1e85be2b272afbce4fb627533a04e6b10dc8b"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.389683 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q" event={"ID":"f23ed77e-a951-4003-bbc2-a2e06431533e","Type":"ContainerStarted","Data":"7348d031f2edbb4ce2dc15b7793fa3c173bbfd7cdb5ac4aeacca99df2096f024"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.390778 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt" event={"ID":"80c510fc-48a8-44c1-9927-c9b0ce934410","Type":"ContainerStarted","Data":"f0346775e298dc70835de1083f9dc5725dc574328dbe8318cb64cfe77ce92f23"} Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.392212 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q" podUID="f23ed77e-a951-4003-bbc2-a2e06431533e" Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.392489 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" event={"ID":"50453e84-c640-4a80-ac7d-0fe2ca24acfd","Type":"ContainerStarted","Data":"a0f4b355cf92263e616560265a6ecadf0b8e9c4a1d2bfa0323e226fb5e7b82cd"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.403547 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" event={"ID":"37b893c7-b6a7-4a19-bc3e-b251deacdd27","Type":"ContainerStarted","Data":"27a6a236d493cd3a82ee7b6725e28c4f7b559acb3f174310fab1d40b10525700"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.406640 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x" event={"ID":"3ae6c6f4-2d25-420b-8149-a976de545cd6","Type":"ContainerStarted","Data":"8c8a8df7f45227659ffd2f675ab815258c93b808797e8730f038e4766f2e284e"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.413852 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h" event={"ID":"e3193e9c-898b-456d-a825-ea8176568803","Type":"ContainerStarted","Data":"9fcb6f81cf2a58e10e5a006de0a6f163fbc30fa3a792c963559ffcf091df9f88"} Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.416282 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr" event={"ID":"0a17b021-e8aa-4f88-b150-596885e48ab1","Type":"ContainerStarted","Data":"ee370af83c5a1de5f7be3b571c0355ac70d22893fbae9bd85bfef57c64f85f01"} Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.455774 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" podUID="d10e063a-f99e-4fc8-b539-224bbf23f466" Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.457629 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" podUID="50453e84-c640-4a80-ac7d-0fe2ca24acfd" Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.543635 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad-cert\") pod \"openstack-operator-controller-manager-b9d6d9595-9zj7r\" (UID: \"26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad\") " pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.550392 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad-cert\") pod \"openstack-operator-controller-manager-b9d6d9595-9zj7r\" (UID: \"26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad\") " pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" Oct 14 15:28:52 crc kubenswrapper[4945]: I1014 15:28:52.569381 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.652919 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" podUID="2c954bb2-bf12-491f-92e8-637e0eb13ad5" Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.656619 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" podUID="d39181a7-852e-402c-9912-7c81a0952e90" Oct 14 15:28:52 crc kubenswrapper[4945]: E1014 15:28:52.656858 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" podUID="2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c" Oct 14 15:28:53 crc kubenswrapper[4945]: I1014 15:28:53.187144 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r"] Oct 14 15:28:53 crc kubenswrapper[4945]: W1014 15:28:53.220834 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26f6d765_fcf8_4f22_a10b_7e9a06a6c0ad.slice/crio-cebd7e9699aa0179118f67e14d34c3c2381a13fc4d0264e873803cb5e1c43a02 WatchSource:0}: Error finding container cebd7e9699aa0179118f67e14d34c3c2381a13fc4d0264e873803cb5e1c43a02: Status 404 returned error can't find the container with id cebd7e9699aa0179118f67e14d34c3c2381a13fc4d0264e873803cb5e1c43a02 Oct 14 15:28:53 crc kubenswrapper[4945]: I1014 15:28:53.478495 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" event={"ID":"2c954bb2-bf12-491f-92e8-637e0eb13ad5","Type":"ContainerStarted","Data":"9430f539595f610d81b4bbfaee4cba994fe4ab2873ce3658e884bdf4d51827ea"} Oct 14 15:28:53 crc kubenswrapper[4945]: E1014 15:28:53.480592 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" podUID="2c954bb2-bf12-491f-92e8-637e0eb13ad5" Oct 14 15:28:53 crc kubenswrapper[4945]: I1014 15:28:53.488095 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" event={"ID":"41d187ec-0816-4800-9d76-9d7a3f364391","Type":"ContainerStarted","Data":"1c8c8139fba84e59c85e98be28edbdf7e70363f1f37e5000b37d20a81608fc1c"} Oct 14 15:28:53 crc kubenswrapper[4945]: I1014 15:28:53.494468 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" event={"ID":"d10e063a-f99e-4fc8-b539-224bbf23f466","Type":"ContainerStarted","Data":"fe52fd2306f401bbefb70ab4840d9c9d39f6ceabd41541f6663401e94673ecd7"} Oct 14 15:28:53 crc kubenswrapper[4945]: E1014 15:28:53.497088 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" podUID="d10e063a-f99e-4fc8-b539-224bbf23f466" Oct 14 15:28:53 crc kubenswrapper[4945]: I1014 15:28:53.497912 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" event={"ID":"669db3c1-3b7e-4bbd-980e-93fdea61a1cc","Type":"ContainerStarted","Data":"12b22294dcbd3962a73306d184246475b17b6777eb6e82aefec26dea9329fc1c"} Oct 14 15:28:53 crc kubenswrapper[4945]: I1014 15:28:53.502533 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" event={"ID":"2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c","Type":"ContainerStarted","Data":"c8e7c9af0a1aae9c259b5897a37ba85431fac345acd4bfe906e64d9acaf99015"} Oct 14 15:28:53 crc kubenswrapper[4945]: E1014 15:28:53.507856 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a\\\"\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" podUID="2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c" Oct 14 15:28:53 crc kubenswrapper[4945]: I1014 15:28:53.542905 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" event={"ID":"50453e84-c640-4a80-ac7d-0fe2ca24acfd","Type":"ContainerStarted","Data":"f0eb88bcd7ce6fde55241eb5999fa7468453294b20f0011bd7a000b53f8cd273"} Oct 14 15:28:53 crc kubenswrapper[4945]: E1014 15:28:53.544181 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" podUID="50453e84-c640-4a80-ac7d-0fe2ca24acfd" Oct 14 15:28:53 crc kubenswrapper[4945]: I1014 15:28:53.570500 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" event={"ID":"d39181a7-852e-402c-9912-7c81a0952e90","Type":"ContainerStarted","Data":"e31330dd8f7c70a1ee6fb74d029ce878c77e50edf78adeb029fa7dfd922a94da"} Oct 14 15:28:53 crc kubenswrapper[4945]: E1014 15:28:53.575625 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" podUID="d39181a7-852e-402c-9912-7c81a0952e90" Oct 14 15:28:53 crc kubenswrapper[4945]: I1014 15:28:53.597099 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" event={"ID":"26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad","Type":"ContainerStarted","Data":"cebd7e9699aa0179118f67e14d34c3c2381a13fc4d0264e873803cb5e1c43a02"} Oct 14 15:28:53 crc kubenswrapper[4945]: E1014 15:28:53.602859 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q" podUID="f23ed77e-a951-4003-bbc2-a2e06431533e" Oct 14 15:28:54 crc kubenswrapper[4945]: I1014 15:28:54.630399 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" event={"ID":"26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad","Type":"ContainerStarted","Data":"d463fe31a304082e7c7e8d7d41ba7c976e13d8ea94c4e7d431fc8351d3eacb28"} Oct 14 15:28:54 crc kubenswrapper[4945]: I1014 15:28:54.630700 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" Oct 14 15:28:54 crc kubenswrapper[4945]: I1014 15:28:54.630714 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" event={"ID":"26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad","Type":"ContainerStarted","Data":"e8247a3a1e8b439b01b55ff3309a3d94cb06509db1ee39f90284a87520604b7a"} Oct 14 15:28:54 crc kubenswrapper[4945]: E1014 15:28:54.632010 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" podUID="d39181a7-852e-402c-9912-7c81a0952e90" Oct 14 15:28:54 crc kubenswrapper[4945]: E1014 15:28:54.632084 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" podUID="2c954bb2-bf12-491f-92e8-637e0eb13ad5" Oct 14 15:28:54 crc kubenswrapper[4945]: E1014 15:28:54.632427 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" podUID="d10e063a-f99e-4fc8-b539-224bbf23f466" Oct 14 15:28:54 crc kubenswrapper[4945]: E1014 15:28:54.632715 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" podUID="50453e84-c640-4a80-ac7d-0fe2ca24acfd" Oct 14 15:28:54 crc kubenswrapper[4945]: E1014 15:28:54.633341 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a\\\"\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" podUID="2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c" Oct 14 15:28:54 crc kubenswrapper[4945]: I1014 15:28:54.728809 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" podStartSLOduration=4.728779863 podStartE2EDuration="4.728779863s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:28:54.720175389 +0000 UTC m=+864.704223757" watchObservedRunningTime="2025-10-14 15:28:54.728779863 +0000 UTC m=+864.712828231" Oct 14 15:29:02 crc kubenswrapper[4945]: I1014 15:29:02.575534 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-b9d6d9595-9zj7r" Oct 14 15:29:06 crc kubenswrapper[4945]: E1014 15:29:06.170048 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:b2e9acf568a48c28cf2aed6012e432eeeb7d5f0eb11878fc91b62bc34cba10cd" Oct 14 15:29:06 crc kubenswrapper[4945]: E1014 15:29:06.170858 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:b2e9acf568a48c28cf2aed6012e432eeeb7d5f0eb11878fc91b62bc34cba10cd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k5f8s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-57bb74c7bf-fl79b_openstack-operators(b4e68062-2195-43f9-a6ae-a58d12bd7ce2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:29:06 crc kubenswrapper[4945]: E1014 15:29:06.919967 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997" Oct 14 15:29:06 crc kubenswrapper[4945]: E1014 15:29:06.920195 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6dhmw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-ddb98f99b-fqb29_openstack-operators(97138e7c-7207-4c47-898f-f71625e0e49b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:29:07 crc kubenswrapper[4945]: E1014 15:29:07.496106 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff" Oct 14 15:29:07 crc kubenswrapper[4945]: E1014 15:29:07.496380 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dpzz9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-664664cb68-58vsj_openstack-operators(01b34d70-1617-4e4d-81ab-8cada4bb1edf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:29:08 crc kubenswrapper[4945]: E1014 15:29:08.186171 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14" Oct 14 15:29:08 crc kubenswrapper[4945]: E1014 15:29:08.186918 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xnckw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-869cc7797f-r6tvw_openstack-operators(37b893c7-b6a7-4a19-bc3e-b251deacdd27): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:29:08 crc kubenswrapper[4945]: E1014 15:29:08.441421 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.72:5001/openstack-k8s-operators/ironic-operator:5a3e546324a5615869429dfb62403ca98d63f158" Oct 14 15:29:08 crc kubenswrapper[4945]: E1014 15:29:08.441512 4945 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.72:5001/openstack-k8s-operators/ironic-operator:5a3e546324a5615869429dfb62403ca98d63f158" Oct 14 15:29:08 crc kubenswrapper[4945]: E1014 15:29:08.441661 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.129.56.72:5001/openstack-k8s-operators/ironic-operator:5a3e546324a5615869429dfb62403ca98d63f158,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pk8h7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-76ccf76b65-p9trl_openstack-operators(20fbc887-9299-460d-9890-f9d7e0235118): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:29:10 crc kubenswrapper[4945]: E1014 15:29:10.050448 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" podUID="b4e68062-2195-43f9-a6ae-a58d12bd7ce2" Oct 14 15:29:10 crc kubenswrapper[4945]: I1014 15:29:10.740929 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" event={"ID":"b4e68062-2195-43f9-a6ae-a58d12bd7ce2","Type":"ContainerStarted","Data":"0f7f5017c8fa2fc99b0e04da4a02e1bb90683dea758f5ef20dbe559f1a19ef15"} Oct 14 15:29:10 crc kubenswrapper[4945]: E1014 15:29:10.745655 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:b2e9acf568a48c28cf2aed6012e432eeeb7d5f0eb11878fc91b62bc34cba10cd\\\"\"" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" podUID="b4e68062-2195-43f9-a6ae-a58d12bd7ce2" Oct 14 15:29:11 crc kubenswrapper[4945]: E1014 15:29:11.737364 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" podUID="37b893c7-b6a7-4a19-bc3e-b251deacdd27" Oct 14 15:29:11 crc kubenswrapper[4945]: I1014 15:29:11.760052 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" event={"ID":"37b893c7-b6a7-4a19-bc3e-b251deacdd27","Type":"ContainerStarted","Data":"9c58055296dd3fba8afbb77284b6cbe4e9dd1d747642787e918a93d2f29317bc"} Oct 14 15:29:11 crc kubenswrapper[4945]: E1014 15:29:11.761844 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:b2e9acf568a48c28cf2aed6012e432eeeb7d5f0eb11878fc91b62bc34cba10cd\\\"\"" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" podUID="b4e68062-2195-43f9-a6ae-a58d12bd7ce2" Oct 14 15:29:11 crc kubenswrapper[4945]: E1014 15:29:11.761914 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" podUID="37b893c7-b6a7-4a19-bc3e-b251deacdd27" Oct 14 15:29:12 crc kubenswrapper[4945]: E1014 15:29:12.882208 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" podUID="37b893c7-b6a7-4a19-bc3e-b251deacdd27" Oct 14 15:29:13 crc kubenswrapper[4945]: E1014 15:29:13.764320 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" podUID="20fbc887-9299-460d-9890-f9d7e0235118" Oct 14 15:29:13 crc kubenswrapper[4945]: E1014 15:29:13.764428 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" podUID="01b34d70-1617-4e4d-81ab-8cada4bb1edf" Oct 14 15:29:13 crc kubenswrapper[4945]: E1014 15:29:13.764495 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" podUID="97138e7c-7207-4c47-898f-f71625e0e49b" Oct 14 15:29:13 crc kubenswrapper[4945]: I1014 15:29:13.780890 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" event={"ID":"01b34d70-1617-4e4d-81ab-8cada4bb1edf","Type":"ContainerStarted","Data":"477453ad0dfbe9af5f8bd93a3936a4e650d5e102dac5cd5a00b4afd1ce139d2e"} Oct 14 15:29:13 crc kubenswrapper[4945]: I1014 15:29:13.782660 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" event={"ID":"20fbc887-9299-460d-9890-f9d7e0235118","Type":"ContainerStarted","Data":"5c14e9c8bc7af30e5bf8441133e7d1f9013a785db00ebb3b592c45dee61dd86f"} Oct 14 15:29:13 crc kubenswrapper[4945]: I1014 15:29:13.784590 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" event={"ID":"97138e7c-7207-4c47-898f-f71625e0e49b","Type":"ContainerStarted","Data":"416b097b625b4a40bbb24a7cf4ad396a55323fd698584fbed32dfe242e76516d"} Oct 14 15:29:14 crc kubenswrapper[4945]: E1014 15:29:14.188393 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.72:5001/openstack-k8s-operators/ironic-operator:5a3e546324a5615869429dfb62403ca98d63f158\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" podUID="20fbc887-9299-460d-9890-f9d7e0235118" Oct 14 15:29:14 crc kubenswrapper[4945]: E1014 15:29:14.189053 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" podUID="01b34d70-1617-4e4d-81ab-8cada4bb1edf" Oct 14 15:29:14 crc kubenswrapper[4945]: E1014 15:29:14.189476 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" podUID="97138e7c-7207-4c47-898f-f71625e0e49b" Oct 14 15:29:14 crc kubenswrapper[4945]: I1014 15:29:14.798796 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6" event={"ID":"f5ac9c23-4d14-460d-9255-16d1865e9aa1","Type":"ContainerStarted","Data":"2f839e841e8ec1ae5931b67c099a5b5c5a9b9b5d353f152a5e48b8e227f84b6d"} Oct 14 15:29:14 crc kubenswrapper[4945]: I1014 15:29:14.808403 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr" event={"ID":"0a17b021-e8aa-4f88-b150-596885e48ab1","Type":"ContainerStarted","Data":"b3f579cc9f31df943711c57601a3da28d4bf81c193547f07736fd0abf80b282f"} Oct 14 15:29:14 crc kubenswrapper[4945]: I1014 15:29:14.809662 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7" event={"ID":"2a652464-b536-4402-a883-f52390e0327b","Type":"ContainerStarted","Data":"56c1bf34030458ece776dfe24c0b3be320b70891ddc6886e7f43fa405821db9a"} Oct 14 15:29:14 crc kubenswrapper[4945]: E1014 15:29:14.926898 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.72:5001/openstack-k8s-operators/ironic-operator:5a3e546324a5615869429dfb62403ca98d63f158\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" podUID="20fbc887-9299-460d-9890-f9d7e0235118" Oct 14 15:29:14 crc kubenswrapper[4945]: E1014 15:29:14.934245 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" podUID="97138e7c-7207-4c47-898f-f71625e0e49b" Oct 14 15:29:14 crc kubenswrapper[4945]: E1014 15:29:14.934301 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" podUID="01b34d70-1617-4e4d-81ab-8cada4bb1edf" Oct 14 15:29:15 crc kubenswrapper[4945]: I1014 15:29:15.817867 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" event={"ID":"669db3c1-3b7e-4bbd-980e-93fdea61a1cc","Type":"ContainerStarted","Data":"136eeda82fca5bb66b2368f71a170f0d818e313faf347e4505db4a7b030b56ea"} Oct 14 15:29:15 crc kubenswrapper[4945]: I1014 15:29:15.819419 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x" event={"ID":"3ae6c6f4-2d25-420b-8149-a976de545cd6","Type":"ContainerStarted","Data":"acb405a1c3cb87f8de9c9af844b6e287eef87f9f80a95f15aee13e34498f41ab"} Oct 14 15:29:15 crc kubenswrapper[4945]: I1014 15:29:15.821488 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j" event={"ID":"86d698df-b089-498a-98e5-65c196aabcc3","Type":"ContainerStarted","Data":"9daf81343424969270836431b602a0e4c6174f3bfd256ab6dca9a6a8af87ad70"} Oct 14 15:29:15 crc kubenswrapper[4945]: I1014 15:29:15.823507 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" event={"ID":"41d187ec-0816-4800-9d76-9d7a3f364391","Type":"ContainerStarted","Data":"d10d140bbd49c84512d2c98a943234ad9aff0164c1f8e7d3893a6d1fa803c539"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.830683 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" event={"ID":"d10e063a-f99e-4fc8-b539-224bbf23f466","Type":"ContainerStarted","Data":"c95d31b051aa3c9e57be95155487a7fad7f39fc656469c84215e59b187bcb7b4"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.830903 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.832961 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" event={"ID":"d39181a7-852e-402c-9912-7c81a0952e90","Type":"ContainerStarted","Data":"79aa5e4b653b6ea11ca232e349179f46d3a6dfe4aa305f9f2e460c33a2fc2789"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.833188 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.834771 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" event={"ID":"41d187ec-0816-4800-9d76-9d7a3f364391","Type":"ContainerStarted","Data":"ac8f0e3bd17476049f8ea09695228f74815fad8be00e5fda5ddd5eda754a26f1"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.837036 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7" event={"ID":"2a652464-b536-4402-a883-f52390e0327b","Type":"ContainerStarted","Data":"9552f4bbd1261dce0b243d058ab1db4bdd66b75451549aec98114ad0ad7bce84"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.837352 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.838925 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x" event={"ID":"3ae6c6f4-2d25-420b-8149-a976de545cd6","Type":"ContainerStarted","Data":"4a802e3a3590431882df032ea3f40ee61701c36bc6868ff6d3b2e3fab1892c63"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.839052 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.840494 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr" event={"ID":"0a17b021-e8aa-4f88-b150-596885e48ab1","Type":"ContainerStarted","Data":"c4ec80e49346d8bf1f5071fc4181a22bcabb68b12fc10715cbde4814a321cfd1"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.840643 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.843069 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t" event={"ID":"64596f58-bc5a-4cdf-80d4-22002c0c9575","Type":"ContainerStarted","Data":"9d81dd611fc54926c6e94f58d51530ceca463c44aa6beb8dc8f31a806ba809c1"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.843210 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t" event={"ID":"64596f58-bc5a-4cdf-80d4-22002c0c9575","Type":"ContainerStarted","Data":"92d16cc858a5506dabd468dd02513ef82e9693fbe67a8112033a8d28153bceb1"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.845167 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" event={"ID":"669db3c1-3b7e-4bbd-980e-93fdea61a1cc","Type":"ContainerStarted","Data":"a08531e16582b896de4141c6e12ee040c7740278ae787ec47a80fe708f9aae91"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.845311 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.847747 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6" event={"ID":"f5ac9c23-4d14-460d-9255-16d1865e9aa1","Type":"ContainerStarted","Data":"3c4c0acb1d50d0230467c4d134796d6b3dfff56fd8ecae11f3582140e6fe92c4"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.847975 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.849547 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h" event={"ID":"e3193e9c-898b-456d-a825-ea8176568803","Type":"ContainerStarted","Data":"8dbc30118c3554ea751691641f7f323b9752b7ad49a03c675e516a82b8be4c89"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.849588 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h" event={"ID":"e3193e9c-898b-456d-a825-ea8176568803","Type":"ContainerStarted","Data":"ffca0de90c93b3d5774dbac7903f992d464def4ed78c7cf441cf3230c00b3dc3"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.851470 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j" event={"ID":"86d698df-b089-498a-98e5-65c196aabcc3","Type":"ContainerStarted","Data":"f0ff1ac357a4467be5d63f2627d0122d5067272b94657cba3b1154c3ed0fba72"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.851785 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.852800 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4" event={"ID":"4af90a70-780d-4d6f-9fc5-6a384002a606","Type":"ContainerStarted","Data":"7466f7cc76474cc9798866d9bb8e2847eed1ff4fa17a7b18564136c2f7921e22"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.854478 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" event={"ID":"2c954bb2-bf12-491f-92e8-637e0eb13ad5","Type":"ContainerStarted","Data":"21c3d0eb183a1b0ef001cb7ecf0b90e46015f91372789c19313128a3d16ba4db"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.855087 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.856718 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q" event={"ID":"f23ed77e-a951-4003-bbc2-a2e06431533e","Type":"ContainerStarted","Data":"8c921bea4c37ebbdc1b4625deac32cc7090855f81e5ea1caf588359e019ad210"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.858914 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" event={"ID":"2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c","Type":"ContainerStarted","Data":"580bbbb8082638629d1dd6743462b631da03127f2009a9a924fba2c39d1b1d6c"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.859289 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.861017 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt" event={"ID":"80c510fc-48a8-44c1-9927-c9b0ce934410","Type":"ContainerStarted","Data":"d36356c0d535dc19d68211edae0ed71e0ecfd4ab4ffedd6f9b0e9521db353dfd"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.865992 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" event={"ID":"50453e84-c640-4a80-ac7d-0fe2ca24acfd","Type":"ContainerStarted","Data":"7bf5f429cb16b25ec2e8b7f551a1ed376603078dcb55193dedf8ea366251a2c5"} Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.866416 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.883251 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" podStartSLOduration=4.134982112 podStartE2EDuration="26.883235982s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:52.048463155 +0000 UTC m=+862.032511523" lastFinishedPulling="2025-10-14 15:29:14.796717025 +0000 UTC m=+884.780765393" observedRunningTime="2025-10-14 15:29:16.881926863 +0000 UTC m=+886.865975231" watchObservedRunningTime="2025-10-14 15:29:16.883235982 +0000 UTC m=+886.867284350" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.926584 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" podStartSLOduration=4.919236965 podStartE2EDuration="26.926563621s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:52.068649781 +0000 UTC m=+862.052698159" lastFinishedPulling="2025-10-14 15:29:14.075976447 +0000 UTC m=+884.060024815" observedRunningTime="2025-10-14 15:29:16.922237693 +0000 UTC m=+886.906286061" watchObservedRunningTime="2025-10-14 15:29:16.926563621 +0000 UTC m=+886.910611989" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.953344 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" podStartSLOduration=4.017189985 podStartE2EDuration="26.953328051s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:52.059155301 +0000 UTC m=+862.043203669" lastFinishedPulling="2025-10-14 15:29:14.995293367 +0000 UTC m=+884.979341735" observedRunningTime="2025-10-14 15:29:16.950095936 +0000 UTC m=+886.934144294" watchObservedRunningTime="2025-10-14 15:29:16.953328051 +0000 UTC m=+886.937376409" Oct 14 15:29:16 crc kubenswrapper[4945]: I1014 15:29:16.969467 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j" podStartSLOduration=10.333749819 podStartE2EDuration="27.969449457s" podCreationTimestamp="2025-10-14 15:28:49 +0000 UTC" firstStartedPulling="2025-10-14 15:28:51.339123294 +0000 UTC m=+861.323171662" lastFinishedPulling="2025-10-14 15:29:08.974822932 +0000 UTC m=+878.958871300" observedRunningTime="2025-10-14 15:29:16.96921765 +0000 UTC m=+886.953266018" watchObservedRunningTime="2025-10-14 15:29:16.969449457 +0000 UTC m=+886.953497825" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.010970 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" podStartSLOduration=9.635556749 podStartE2EDuration="27.010951252s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:52.342437754 +0000 UTC m=+862.326486122" lastFinishedPulling="2025-10-14 15:29:09.717832257 +0000 UTC m=+879.701880625" observedRunningTime="2025-10-14 15:29:17.008638924 +0000 UTC m=+886.992687312" watchObservedRunningTime="2025-10-14 15:29:17.010951252 +0000 UTC m=+886.994999620" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.034212 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7" podStartSLOduration=9.731025878 podStartE2EDuration="27.034198159s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:51.681002467 +0000 UTC m=+861.665050835" lastFinishedPulling="2025-10-14 15:29:08.984174748 +0000 UTC m=+878.968223116" observedRunningTime="2025-10-14 15:29:17.030928542 +0000 UTC m=+887.014976930" watchObservedRunningTime="2025-10-14 15:29:17.034198159 +0000 UTC m=+887.018246527" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.051880 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x" podStartSLOduration=10.690977095000001 podStartE2EDuration="28.05186061s" podCreationTimestamp="2025-10-14 15:28:49 +0000 UTC" firstStartedPulling="2025-10-14 15:28:51.623010885 +0000 UTC m=+861.607059253" lastFinishedPulling="2025-10-14 15:29:08.9838944 +0000 UTC m=+878.967942768" observedRunningTime="2025-10-14 15:29:17.049901102 +0000 UTC m=+887.033949470" watchObservedRunningTime="2025-10-14 15:29:17.05186061 +0000 UTC m=+887.035908978" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.074485 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" podStartSLOduration=6.993896033 podStartE2EDuration="27.074468848s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:52.07303 +0000 UTC m=+862.057078368" lastFinishedPulling="2025-10-14 15:29:12.153602815 +0000 UTC m=+882.137651183" observedRunningTime="2025-10-14 15:29:17.073035295 +0000 UTC m=+887.057083663" watchObservedRunningTime="2025-10-14 15:29:17.074468848 +0000 UTC m=+887.058517216" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.117103 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" podStartSLOduration=4.383953643 podStartE2EDuration="27.117088546s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:52.065147288 +0000 UTC m=+862.049195656" lastFinishedPulling="2025-10-14 15:29:14.798282191 +0000 UTC m=+884.782330559" observedRunningTime="2025-10-14 15:29:17.09655857 +0000 UTC m=+887.080606958" watchObservedRunningTime="2025-10-14 15:29:17.117088546 +0000 UTC m=+887.101136914" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.118400 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6" podStartSLOduration=10.720227088 podStartE2EDuration="28.118393464s" podCreationTimestamp="2025-10-14 15:28:49 +0000 UTC" firstStartedPulling="2025-10-14 15:28:50.943845855 +0000 UTC m=+860.927894223" lastFinishedPulling="2025-10-14 15:29:08.342012231 +0000 UTC m=+878.326060599" observedRunningTime="2025-10-14 15:29:17.116060016 +0000 UTC m=+887.100108384" watchObservedRunningTime="2025-10-14 15:29:17.118393464 +0000 UTC m=+887.102441832" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.139914 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr" podStartSLOduration=10.772283416 podStartE2EDuration="28.139896769s" podCreationTimestamp="2025-10-14 15:28:49 +0000 UTC" firstStartedPulling="2025-10-14 15:28:51.607622281 +0000 UTC m=+861.591670659" lastFinishedPulling="2025-10-14 15:29:08.975235644 +0000 UTC m=+878.959284012" observedRunningTime="2025-10-14 15:29:17.133910852 +0000 UTC m=+887.117959220" watchObservedRunningTime="2025-10-14 15:29:17.139896769 +0000 UTC m=+887.123945137" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.154701 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q" podStartSLOduration=7.05104667 podStartE2EDuration="27.154680136s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:52.048504756 +0000 UTC m=+862.032553124" lastFinishedPulling="2025-10-14 15:29:12.152138222 +0000 UTC m=+882.136186590" observedRunningTime="2025-10-14 15:29:17.150394989 +0000 UTC m=+887.134443357" watchObservedRunningTime="2025-10-14 15:29:17.154680136 +0000 UTC m=+887.138728504" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.881649 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4" event={"ID":"4af90a70-780d-4d6f-9fc5-6a384002a606","Type":"ContainerStarted","Data":"a1b2fa0c97f6356ab3b94b872aceacdcce2d256cead0834e0fc1d13639bca173"} Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.882142 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.887076 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt" event={"ID":"80c510fc-48a8-44c1-9927-c9b0ce934410","Type":"ContainerStarted","Data":"191336e4dd7f248e74ba7b6e0c6ccfd07433b6d6ed09a29bdba6275800fb2504"} Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.888413 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.902827 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4" podStartSLOduration=10.96880823 podStartE2EDuration="27.902806482s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:52.047774755 +0000 UTC m=+862.031823113" lastFinishedPulling="2025-10-14 15:29:08.981772997 +0000 UTC m=+878.965821365" observedRunningTime="2025-10-14 15:29:17.896603969 +0000 UTC m=+887.880652377" watchObservedRunningTime="2025-10-14 15:29:17.902806482 +0000 UTC m=+887.886854860" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.918174 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h" podStartSLOduration=10.967146761 podStartE2EDuration="27.918150865s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:52.024165118 +0000 UTC m=+862.008213486" lastFinishedPulling="2025-10-14 15:29:08.975169222 +0000 UTC m=+878.959217590" observedRunningTime="2025-10-14 15:29:17.911088356 +0000 UTC m=+887.895136724" watchObservedRunningTime="2025-10-14 15:29:17.918150865 +0000 UTC m=+887.902199233" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.932343 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" podStartSLOduration=12.279159262 podStartE2EDuration="28.932318763s" podCreationTimestamp="2025-10-14 15:28:49 +0000 UTC" firstStartedPulling="2025-10-14 15:28:52.321769914 +0000 UTC m=+862.305818282" lastFinishedPulling="2025-10-14 15:29:08.974929405 +0000 UTC m=+878.958977783" observedRunningTime="2025-10-14 15:29:17.931303383 +0000 UTC m=+887.915351771" watchObservedRunningTime="2025-10-14 15:29:17.932318763 +0000 UTC m=+887.916367151" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.951420 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t" podStartSLOduration=11.279647754 podStartE2EDuration="28.951394956s" podCreationTimestamp="2025-10-14 15:28:49 +0000 UTC" firstStartedPulling="2025-10-14 15:28:51.303550104 +0000 UTC m=+861.287598472" lastFinishedPulling="2025-10-14 15:29:08.975297306 +0000 UTC m=+878.959345674" observedRunningTime="2025-10-14 15:29:17.948424959 +0000 UTC m=+887.932473327" watchObservedRunningTime="2025-10-14 15:29:17.951394956 +0000 UTC m=+887.935443334" Oct 14 15:29:17 crc kubenswrapper[4945]: I1014 15:29:17.965746 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt" podStartSLOduration=10.742840937 podStartE2EDuration="28.965727359s" podCreationTimestamp="2025-10-14 15:28:49 +0000 UTC" firstStartedPulling="2025-10-14 15:28:51.625811618 +0000 UTC m=+861.609859986" lastFinishedPulling="2025-10-14 15:29:09.84869804 +0000 UTC m=+879.832746408" observedRunningTime="2025-10-14 15:29:17.962465403 +0000 UTC m=+887.946513771" watchObservedRunningTime="2025-10-14 15:29:17.965727359 +0000 UTC m=+887.949775717" Oct 14 15:29:18 crc kubenswrapper[4945]: I1014 15:29:18.892691 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt" Oct 14 15:29:20 crc kubenswrapper[4945]: I1014 15:29:20.205991 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xlwn6" Oct 14 15:29:20 crc kubenswrapper[4945]: I1014 15:29:20.229947 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-6fg2j" Oct 14 15:29:20 crc kubenswrapper[4945]: I1014 15:29:20.240343 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t" Oct 14 15:29:20 crc kubenswrapper[4945]: I1014 15:29:20.280258 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-8gm4x" Oct 14 15:29:20 crc kubenswrapper[4945]: I1014 15:29:20.318608 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-v9hhr" Oct 14 15:29:20 crc kubenswrapper[4945]: I1014 15:29:20.634060 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-59578bc799-jcrs7" Oct 14 15:29:21 crc kubenswrapper[4945]: I1014 15:29:21.025732 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-kkqht" Oct 14 15:29:21 crc kubenswrapper[4945]: I1014 15:29:21.045913 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjvtx" Oct 14 15:29:21 crc kubenswrapper[4945]: I1014 15:29:21.067946 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w" Oct 14 15:29:21 crc kubenswrapper[4945]: I1014 15:29:21.102165 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" Oct 14 15:29:21 crc kubenswrapper[4945]: I1014 15:29:21.119041 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-pk25j" Oct 14 15:29:30 crc kubenswrapper[4945]: I1014 15:29:30.242335 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-gf66t" Oct 14 15:29:30 crc kubenswrapper[4945]: I1014 15:29:30.327574 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-wchqt" Oct 14 15:29:30 crc kubenswrapper[4945]: I1014 15:29:30.622556 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-h2p64" Oct 14 15:29:30 crc kubenswrapper[4945]: I1014 15:29:30.652038 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-qx6hl" Oct 14 15:29:30 crc kubenswrapper[4945]: I1014 15:29:30.660341 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-48x4h" Oct 14 15:29:30 crc kubenswrapper[4945]: I1014 15:29:30.692967 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-48769" Oct 14 15:29:30 crc kubenswrapper[4945]: I1014 15:29:30.726253 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-vspm4" Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.005817 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" event={"ID":"01b34d70-1617-4e4d-81ab-8cada4bb1edf","Type":"ContainerStarted","Data":"4551ffa4dce6dd871d7a960abfac33372c2bd07137f019bc505f8c5796d4b7d1"} Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.006575 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.007580 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" event={"ID":"20fbc887-9299-460d-9890-f9d7e0235118","Type":"ContainerStarted","Data":"70a424d13548063deeebeaba1d7cf93676c9a0a515c59dced533e5f86e29482e"} Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.007723 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.010057 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" event={"ID":"b4e68062-2195-43f9-a6ae-a58d12bd7ce2","Type":"ContainerStarted","Data":"6d0ea306c0341c23ca247c98793a546fd17519f73760dcb4795f92a900efbbc6"} Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.010222 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.012025 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" event={"ID":"97138e7c-7207-4c47-898f-f71625e0e49b","Type":"ContainerStarted","Data":"d9727396b8ad8e4823ae6f27d1804b61f42d98e4787a1141d7f6de937fe04283"} Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.012289 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.013737 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" event={"ID":"37b893c7-b6a7-4a19-bc3e-b251deacdd27","Type":"ContainerStarted","Data":"2dbb436a9457efa0b07763f167a8f5da99c89e2decd61a041a5bb0c339df4495"} Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.013950 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.034136 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" podStartSLOduration=3.252857611 podStartE2EDuration="45.034116259s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:52.025063824 +0000 UTC m=+862.009112192" lastFinishedPulling="2025-10-14 15:29:33.806322472 +0000 UTC m=+903.790370840" observedRunningTime="2025-10-14 15:29:35.030397179 +0000 UTC m=+905.014445577" watchObservedRunningTime="2025-10-14 15:29:35.034116259 +0000 UTC m=+905.018164627" Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.048945 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" podStartSLOduration=3.942885601 podStartE2EDuration="46.048926436s" podCreationTimestamp="2025-10-14 15:28:49 +0000 UTC" firstStartedPulling="2025-10-14 15:28:51.694378662 +0000 UTC m=+861.678427030" lastFinishedPulling="2025-10-14 15:29:33.800419487 +0000 UTC m=+903.784467865" observedRunningTime="2025-10-14 15:29:35.045888136 +0000 UTC m=+905.029936504" watchObservedRunningTime="2025-10-14 15:29:35.048926436 +0000 UTC m=+905.032974804" Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.061105 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" podStartSLOduration=3.893470162 podStartE2EDuration="46.061087245s" podCreationTimestamp="2025-10-14 15:28:49 +0000 UTC" firstStartedPulling="2025-10-14 15:28:51.633411222 +0000 UTC m=+861.617459590" lastFinishedPulling="2025-10-14 15:29:33.801028295 +0000 UTC m=+903.785076673" observedRunningTime="2025-10-14 15:29:35.061001873 +0000 UTC m=+905.045050261" watchObservedRunningTime="2025-10-14 15:29:35.061087245 +0000 UTC m=+905.045135613" Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.082021 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" podStartSLOduration=2.976120221 podStartE2EDuration="45.082001002s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:51.691472736 +0000 UTC m=+861.675521104" lastFinishedPulling="2025-10-14 15:29:33.797353517 +0000 UTC m=+903.781401885" observedRunningTime="2025-10-14 15:29:35.076801529 +0000 UTC m=+905.060849887" watchObservedRunningTime="2025-10-14 15:29:35.082001002 +0000 UTC m=+905.066049370" Oct 14 15:29:35 crc kubenswrapper[4945]: I1014 15:29:35.093778 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" podStartSLOduration=2.9037170039999998 podStartE2EDuration="45.09375913s" podCreationTimestamp="2025-10-14 15:28:50 +0000 UTC" firstStartedPulling="2025-10-14 15:28:51.607403964 +0000 UTC m=+861.591452332" lastFinishedPulling="2025-10-14 15:29:33.79744608 +0000 UTC m=+903.781494458" observedRunningTime="2025-10-14 15:29:35.090639037 +0000 UTC m=+905.074687405" watchObservedRunningTime="2025-10-14 15:29:35.09375913 +0000 UTC m=+905.077807498" Oct 14 15:29:40 crc kubenswrapper[4945]: I1014 15:29:40.406079 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-fl79b" Oct 14 15:29:40 crc kubenswrapper[4945]: I1014 15:29:40.489471 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-76ccf76b65-p9trl" Oct 14 15:29:40 crc kubenswrapper[4945]: I1014 15:29:40.525176 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-fqb29" Oct 14 15:29:40 crc kubenswrapper[4945]: I1014 15:29:40.600961 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-r6tvw" Oct 14 15:29:40 crc kubenswrapper[4945]: I1014 15:29:40.625757 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-664664cb68-58vsj" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.467651 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-pttvf"] Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.469469 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.472745 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.473108 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-gjdnf" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.473201 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.474393 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.487387 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-pttvf"] Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.516581 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5whp9\" (UniqueName: \"kubernetes.io/projected/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48-kube-api-access-5whp9\") pod \"dnsmasq-dns-675f4bcbfc-pttvf\" (UID: \"9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48\") " pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.516779 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48-config\") pod \"dnsmasq-dns-675f4bcbfc-pttvf\" (UID: \"9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48\") " pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.543584 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2qkbt"] Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.547042 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.551498 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.551656 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2qkbt"] Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.618038 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48-config\") pod \"dnsmasq-dns-675f4bcbfc-pttvf\" (UID: \"9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48\") " pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.618104 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40d751f7-802d-41ee-9afb-16f33e49f0c0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2qkbt\" (UID: \"40d751f7-802d-41ee-9afb-16f33e49f0c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.618140 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5whp9\" (UniqueName: \"kubernetes.io/projected/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48-kube-api-access-5whp9\") pod \"dnsmasq-dns-675f4bcbfc-pttvf\" (UID: \"9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48\") " pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.618164 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw7xj\" (UniqueName: \"kubernetes.io/projected/40d751f7-802d-41ee-9afb-16f33e49f0c0-kube-api-access-mw7xj\") pod \"dnsmasq-dns-78dd6ddcc-2qkbt\" (UID: \"40d751f7-802d-41ee-9afb-16f33e49f0c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.618189 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40d751f7-802d-41ee-9afb-16f33e49f0c0-config\") pod \"dnsmasq-dns-78dd6ddcc-2qkbt\" (UID: \"40d751f7-802d-41ee-9afb-16f33e49f0c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.619124 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48-config\") pod \"dnsmasq-dns-675f4bcbfc-pttvf\" (UID: \"9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48\") " pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.652646 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5whp9\" (UniqueName: \"kubernetes.io/projected/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48-kube-api-access-5whp9\") pod \"dnsmasq-dns-675f4bcbfc-pttvf\" (UID: \"9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48\") " pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.719453 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40d751f7-802d-41ee-9afb-16f33e49f0c0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2qkbt\" (UID: \"40d751f7-802d-41ee-9afb-16f33e49f0c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.719811 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw7xj\" (UniqueName: \"kubernetes.io/projected/40d751f7-802d-41ee-9afb-16f33e49f0c0-kube-api-access-mw7xj\") pod \"dnsmasq-dns-78dd6ddcc-2qkbt\" (UID: \"40d751f7-802d-41ee-9afb-16f33e49f0c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.720194 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40d751f7-802d-41ee-9afb-16f33e49f0c0-config\") pod \"dnsmasq-dns-78dd6ddcc-2qkbt\" (UID: \"40d751f7-802d-41ee-9afb-16f33e49f0c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.720576 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40d751f7-802d-41ee-9afb-16f33e49f0c0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2qkbt\" (UID: \"40d751f7-802d-41ee-9afb-16f33e49f0c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.720952 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40d751f7-802d-41ee-9afb-16f33e49f0c0-config\") pod \"dnsmasq-dns-78dd6ddcc-2qkbt\" (UID: \"40d751f7-802d-41ee-9afb-16f33e49f0c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.745055 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw7xj\" (UniqueName: \"kubernetes.io/projected/40d751f7-802d-41ee-9afb-16f33e49f0c0-kube-api-access-mw7xj\") pod \"dnsmasq-dns-78dd6ddcc-2qkbt\" (UID: \"40d751f7-802d-41ee-9afb-16f33e49f0c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.798782 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" Oct 14 15:29:56 crc kubenswrapper[4945]: I1014 15:29:56.870101 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:29:57 crc kubenswrapper[4945]: I1014 15:29:57.224080 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-pttvf"] Oct 14 15:29:57 crc kubenswrapper[4945]: I1014 15:29:57.290605 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2qkbt"] Oct 14 15:29:57 crc kubenswrapper[4945]: W1014 15:29:57.296251 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40d751f7_802d_41ee_9afb_16f33e49f0c0.slice/crio-127ae298f6900314d6ae84df2e70dc3e5187b4e1a9006f0b690a7c95c73a433f WatchSource:0}: Error finding container 127ae298f6900314d6ae84df2e70dc3e5187b4e1a9006f0b690a7c95c73a433f: Status 404 returned error can't find the container with id 127ae298f6900314d6ae84df2e70dc3e5187b4e1a9006f0b690a7c95c73a433f Oct 14 15:29:58 crc kubenswrapper[4945]: I1014 15:29:58.166966 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" event={"ID":"40d751f7-802d-41ee-9afb-16f33e49f0c0","Type":"ContainerStarted","Data":"127ae298f6900314d6ae84df2e70dc3e5187b4e1a9006f0b690a7c95c73a433f"} Oct 14 15:29:58 crc kubenswrapper[4945]: I1014 15:29:58.170083 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" event={"ID":"9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48","Type":"ContainerStarted","Data":"f421bb05563e381280a4fb7174f83c4bf11bf5cb1f6ba3546751c1437223048e"} Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.532821 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-pttvf"] Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.558359 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-pxxrt"] Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.563546 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.571355 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-pxxrt"] Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.660377 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-config\") pod \"dnsmasq-dns-666b6646f7-pxxrt\" (UID: \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\") " pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.660511 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-pxxrt\" (UID: \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\") " pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.660592 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppzzv\" (UniqueName: \"kubernetes.io/projected/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-kube-api-access-ppzzv\") pod \"dnsmasq-dns-666b6646f7-pxxrt\" (UID: \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\") " pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.761767 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-config\") pod \"dnsmasq-dns-666b6646f7-pxxrt\" (UID: \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\") " pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.762970 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-config\") pod \"dnsmasq-dns-666b6646f7-pxxrt\" (UID: \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\") " pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.763852 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-pxxrt\" (UID: \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\") " pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.763080 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-pxxrt\" (UID: \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\") " pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.763992 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppzzv\" (UniqueName: \"kubernetes.io/projected/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-kube-api-access-ppzzv\") pod \"dnsmasq-dns-666b6646f7-pxxrt\" (UID: \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\") " pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.790346 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppzzv\" (UniqueName: \"kubernetes.io/projected/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-kube-api-access-ppzzv\") pod \"dnsmasq-dns-666b6646f7-pxxrt\" (UID: \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\") " pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.843479 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2qkbt"] Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.870115 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mfqx8"] Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.871535 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.878992 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mfqx8"] Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.893448 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.969253 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39764f61-839c-4425-9b8b-13e92865b7d2-config\") pod \"dnsmasq-dns-57d769cc4f-mfqx8\" (UID: \"39764f61-839c-4425-9b8b-13e92865b7d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.969336 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhfgw\" (UniqueName: \"kubernetes.io/projected/39764f61-839c-4425-9b8b-13e92865b7d2-kube-api-access-mhfgw\") pod \"dnsmasq-dns-57d769cc4f-mfqx8\" (UID: \"39764f61-839c-4425-9b8b-13e92865b7d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:29:59 crc kubenswrapper[4945]: I1014 15:29:59.969371 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39764f61-839c-4425-9b8b-13e92865b7d2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mfqx8\" (UID: \"39764f61-839c-4425-9b8b-13e92865b7d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.071008 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhfgw\" (UniqueName: \"kubernetes.io/projected/39764f61-839c-4425-9b8b-13e92865b7d2-kube-api-access-mhfgw\") pod \"dnsmasq-dns-57d769cc4f-mfqx8\" (UID: \"39764f61-839c-4425-9b8b-13e92865b7d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.071327 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39764f61-839c-4425-9b8b-13e92865b7d2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mfqx8\" (UID: \"39764f61-839c-4425-9b8b-13e92865b7d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.071394 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39764f61-839c-4425-9b8b-13e92865b7d2-config\") pod \"dnsmasq-dns-57d769cc4f-mfqx8\" (UID: \"39764f61-839c-4425-9b8b-13e92865b7d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.072938 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39764f61-839c-4425-9b8b-13e92865b7d2-config\") pod \"dnsmasq-dns-57d769cc4f-mfqx8\" (UID: \"39764f61-839c-4425-9b8b-13e92865b7d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.073338 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39764f61-839c-4425-9b8b-13e92865b7d2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mfqx8\" (UID: \"39764f61-839c-4425-9b8b-13e92865b7d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.094701 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhfgw\" (UniqueName: \"kubernetes.io/projected/39764f61-839c-4425-9b8b-13e92865b7d2-kube-api-access-mhfgw\") pod \"dnsmasq-dns-57d769cc4f-mfqx8\" (UID: \"39764f61-839c-4425-9b8b-13e92865b7d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.140335 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b"] Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.141491 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.143961 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.143961 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.164144 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b"] Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.196906 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.274953 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a85b7c5c-09cb-413d-a321-18aaa9728a52-config-volume\") pod \"collect-profiles-29340930-q7k7b\" (UID: \"a85b7c5c-09cb-413d-a321-18aaa9728a52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.275024 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j997v\" (UniqueName: \"kubernetes.io/projected/a85b7c5c-09cb-413d-a321-18aaa9728a52-kube-api-access-j997v\") pod \"collect-profiles-29340930-q7k7b\" (UID: \"a85b7c5c-09cb-413d-a321-18aaa9728a52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.275058 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a85b7c5c-09cb-413d-a321-18aaa9728a52-secret-volume\") pod \"collect-profiles-29340930-q7k7b\" (UID: \"a85b7c5c-09cb-413d-a321-18aaa9728a52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.377078 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a85b7c5c-09cb-413d-a321-18aaa9728a52-config-volume\") pod \"collect-profiles-29340930-q7k7b\" (UID: \"a85b7c5c-09cb-413d-a321-18aaa9728a52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.377153 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j997v\" (UniqueName: \"kubernetes.io/projected/a85b7c5c-09cb-413d-a321-18aaa9728a52-kube-api-access-j997v\") pod \"collect-profiles-29340930-q7k7b\" (UID: \"a85b7c5c-09cb-413d-a321-18aaa9728a52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.377186 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a85b7c5c-09cb-413d-a321-18aaa9728a52-secret-volume\") pod \"collect-profiles-29340930-q7k7b\" (UID: \"a85b7c5c-09cb-413d-a321-18aaa9728a52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.377903 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a85b7c5c-09cb-413d-a321-18aaa9728a52-config-volume\") pod \"collect-profiles-29340930-q7k7b\" (UID: \"a85b7c5c-09cb-413d-a321-18aaa9728a52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.385316 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a85b7c5c-09cb-413d-a321-18aaa9728a52-secret-volume\") pod \"collect-profiles-29340930-q7k7b\" (UID: \"a85b7c5c-09cb-413d-a321-18aaa9728a52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.412856 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j997v\" (UniqueName: \"kubernetes.io/projected/a85b7c5c-09cb-413d-a321-18aaa9728a52-kube-api-access-j997v\") pod \"collect-profiles-29340930-q7k7b\" (UID: \"a85b7c5c-09cb-413d-a321-18aaa9728a52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.470072 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-pxxrt"] Oct 14 15:30:00 crc kubenswrapper[4945]: W1014 15:30:00.473504 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea8fbc4f_f6ee_405b_9cf3_e88cce3f5b8c.slice/crio-b8a4b4c3ab87d54e2aea7e733ddc10e4b9602d8bd5b8130527b60d655b4a71a7 WatchSource:0}: Error finding container b8a4b4c3ab87d54e2aea7e733ddc10e4b9602d8bd5b8130527b60d655b4a71a7: Status 404 returned error can't find the container with id b8a4b4c3ab87d54e2aea7e733ddc10e4b9602d8bd5b8130527b60d655b4a71a7 Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.479015 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.699988 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.701309 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.705446 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.706488 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-hgnp5" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.708583 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.708800 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.709220 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.709515 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.709697 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.711796 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.756444 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mfqx8"] Oct 14 15:30:00 crc kubenswrapper[4945]: W1014 15:30:00.771859 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39764f61_839c_4425_9b8b_13e92865b7d2.slice/crio-ea6fdfc8b1a1f31e5aab49d6a135b032bdc2a9fea7835d1283911f56990a79cc WatchSource:0}: Error finding container ea6fdfc8b1a1f31e5aab49d6a135b032bdc2a9fea7835d1283911f56990a79cc: Status 404 returned error can't find the container with id ea6fdfc8b1a1f31e5aab49d6a135b032bdc2a9fea7835d1283911f56990a79cc Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.885305 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2spth\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-kube-api-access-2spth\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.885370 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.885533 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.885630 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c7915803-bfbc-4150-854d-6039c685a771-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.885661 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.885778 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.885815 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.885888 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.885914 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-config-data\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.885938 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.885965 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c7915803-bfbc-4150-854d-6039c685a771-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.935228 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b"] Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.991527 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2spth\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-kube-api-access-2spth\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.991593 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.991627 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.991646 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c7915803-bfbc-4150-854d-6039c685a771-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.991665 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.992536 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.992575 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.992595 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.992623 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.992641 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.992679 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-config-data\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.992698 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c7915803-bfbc-4150-854d-6039c685a771-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.994971 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.995175 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.997378 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.997920 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.998090 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-config-data\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.998125 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:00 crc kubenswrapper[4945]: I1014 15:30:00.999668 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.000128 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c7915803-bfbc-4150-854d-6039c685a771-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.001311 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.004687 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c7915803-bfbc-4150-854d-6039c685a771-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.008566 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.008930 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.009049 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.009269 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qpstp" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.009420 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.010524 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.010961 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.012818 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2spth\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-kube-api-access-2spth\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.015287 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.079928 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.082581 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " pod="openstack/rabbitmq-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.093522 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9642a3f8-eb9b-4f79-8448-602a5f7761d6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.093651 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.094016 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.094105 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.094190 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.094224 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.094251 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.094323 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.094461 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9642a3f8-eb9b-4f79-8448-602a5f7761d6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.094517 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.094712 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q2n2\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-kube-api-access-8q2n2\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.195717 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.196093 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.196100 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.196139 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.196194 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.196240 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.196276 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.196313 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.196335 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9642a3f8-eb9b-4f79-8448-602a5f7761d6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.196357 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.196399 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q2n2\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-kube-api-access-8q2n2\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.196421 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9642a3f8-eb9b-4f79-8448-602a5f7761d6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.197079 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.197144 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.197288 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.197808 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.199420 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.201511 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" event={"ID":"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c","Type":"ContainerStarted","Data":"b8a4b4c3ab87d54e2aea7e733ddc10e4b9602d8bd5b8130527b60d655b4a71a7"} Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.203504 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9642a3f8-eb9b-4f79-8448-602a5f7761d6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.206634 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.207251 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.209855 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9642a3f8-eb9b-4f79-8448-602a5f7761d6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.212203 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" event={"ID":"a85b7c5c-09cb-413d-a321-18aaa9728a52","Type":"ContainerStarted","Data":"64cb2a088ad71d4f8a647bc2cfc4d99f61b4792f798a59a8650e9e56d2741d2b"} Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.213708 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" event={"ID":"39764f61-839c-4425-9b8b-13e92865b7d2","Type":"ContainerStarted","Data":"ea6fdfc8b1a1f31e5aab49d6a135b032bdc2a9fea7835d1283911f56990a79cc"} Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.220398 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q2n2\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-kube-api-access-8q2n2\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.231295 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.335474 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 15:30:01 crc kubenswrapper[4945]: I1014 15:30:01.463256 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.226064 4945 generic.go:334] "Generic (PLEG): container finished" podID="a85b7c5c-09cb-413d-a321-18aaa9728a52" containerID="6407c7307acc2af5e39e44e6eb4ec51e1156f7b281375da7c19d463a665187e5" exitCode=0 Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.226117 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" event={"ID":"a85b7c5c-09cb-413d-a321-18aaa9728a52","Type":"ContainerDied","Data":"6407c7307acc2af5e39e44e6eb4ec51e1156f7b281375da7c19d463a665187e5"} Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.611160 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.613831 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.623650 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.624363 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-9bknb" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.625116 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.625228 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.625931 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.626693 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.628574 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.726579 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/afed9c20-1964-44dd-a849-af9fc0f97e86-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.726633 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/afed9c20-1964-44dd-a849-af9fc0f97e86-config-data-default\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.726661 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/afed9c20-1964-44dd-a849-af9fc0f97e86-config-data-generated\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.726776 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.726864 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/afed9c20-1964-44dd-a849-af9fc0f97e86-secrets\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.726914 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afed9c20-1964-44dd-a849-af9fc0f97e86-operator-scripts\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.726984 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd9fl\" (UniqueName: \"kubernetes.io/projected/afed9c20-1964-44dd-a849-af9fc0f97e86-kube-api-access-bd9fl\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.727034 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afed9c20-1964-44dd-a849-af9fc0f97e86-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.727053 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/afed9c20-1964-44dd-a849-af9fc0f97e86-kolla-config\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.828765 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd9fl\" (UniqueName: \"kubernetes.io/projected/afed9c20-1964-44dd-a849-af9fc0f97e86-kube-api-access-bd9fl\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.828827 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afed9c20-1964-44dd-a849-af9fc0f97e86-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.828845 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/afed9c20-1964-44dd-a849-af9fc0f97e86-kolla-config\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.828923 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/afed9c20-1964-44dd-a849-af9fc0f97e86-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.828944 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/afed9c20-1964-44dd-a849-af9fc0f97e86-config-data-default\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.828969 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/afed9c20-1964-44dd-a849-af9fc0f97e86-config-data-generated\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.828988 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.829013 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/afed9c20-1964-44dd-a849-af9fc0f97e86-secrets\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.829030 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afed9c20-1964-44dd-a849-af9fc0f97e86-operator-scripts\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.829380 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.830097 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/afed9c20-1964-44dd-a849-af9fc0f97e86-config-data-generated\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.831406 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/afed9c20-1964-44dd-a849-af9fc0f97e86-kolla-config\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.831964 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/afed9c20-1964-44dd-a849-af9fc0f97e86-config-data-default\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.833830 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afed9c20-1964-44dd-a849-af9fc0f97e86-operator-scripts\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.835291 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/afed9c20-1964-44dd-a849-af9fc0f97e86-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.849193 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd9fl\" (UniqueName: \"kubernetes.io/projected/afed9c20-1964-44dd-a849-af9fc0f97e86-kube-api-access-bd9fl\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.851716 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/afed9c20-1964-44dd-a849-af9fc0f97e86-secrets\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.859761 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.874193 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afed9c20-1964-44dd-a849-af9fc0f97e86-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"afed9c20-1964-44dd-a849-af9fc0f97e86\") " pod="openstack/openstack-galera-0" Oct 14 15:30:02 crc kubenswrapper[4945]: I1014 15:30:02.946572 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.062630 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.064131 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.072311 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.072576 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-5hw4z" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.072676 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.073000 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.080344 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.149136 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-498dl\" (UniqueName: \"kubernetes.io/projected/f0fa2603-a8a2-449d-88d2-e35975a18ad1-kube-api-access-498dl\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.149218 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fa2603-a8a2-449d-88d2-e35975a18ad1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.149287 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0fa2603-a8a2-449d-88d2-e35975a18ad1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.149318 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.149343 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f0fa2603-a8a2-449d-88d2-e35975a18ad1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.149364 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0fa2603-a8a2-449d-88d2-e35975a18ad1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.149394 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f0fa2603-a8a2-449d-88d2-e35975a18ad1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.149425 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f0fa2603-a8a2-449d-88d2-e35975a18ad1-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.149457 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f0fa2603-a8a2-449d-88d2-e35975a18ad1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.234249 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.235497 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.237149 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.241334 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-cw8ws" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.241568 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.242074 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251300 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pktm9\" (UniqueName: \"kubernetes.io/projected/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-kube-api-access-pktm9\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251386 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-498dl\" (UniqueName: \"kubernetes.io/projected/f0fa2603-a8a2-449d-88d2-e35975a18ad1-kube-api-access-498dl\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-kolla-config\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251448 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-config-data\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251476 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fa2603-a8a2-449d-88d2-e35975a18ad1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251506 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-combined-ca-bundle\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251557 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-memcached-tls-certs\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251586 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0fa2603-a8a2-449d-88d2-e35975a18ad1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251616 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251640 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f0fa2603-a8a2-449d-88d2-e35975a18ad1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251675 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0fa2603-a8a2-449d-88d2-e35975a18ad1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251702 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f0fa2603-a8a2-449d-88d2-e35975a18ad1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251726 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f0fa2603-a8a2-449d-88d2-e35975a18ad1-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.251758 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f0fa2603-a8a2-449d-88d2-e35975a18ad1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.252730 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f0fa2603-a8a2-449d-88d2-e35975a18ad1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.252918 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f0fa2603-a8a2-449d-88d2-e35975a18ad1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.253661 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fa2603-a8a2-449d-88d2-e35975a18ad1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.253849 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.254098 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f0fa2603-a8a2-449d-88d2-e35975a18ad1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.262010 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0fa2603-a8a2-449d-88d2-e35975a18ad1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.266291 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f0fa2603-a8a2-449d-88d2-e35975a18ad1-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.270818 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0fa2603-a8a2-449d-88d2-e35975a18ad1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.280640 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-498dl\" (UniqueName: \"kubernetes.io/projected/f0fa2603-a8a2-449d-88d2-e35975a18ad1-kube-api-access-498dl\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.288976 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f0fa2603-a8a2-449d-88d2-e35975a18ad1\") " pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.353311 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-combined-ca-bundle\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.353400 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-memcached-tls-certs\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.353489 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pktm9\" (UniqueName: \"kubernetes.io/projected/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-kube-api-access-pktm9\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.353534 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-kolla-config\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.353554 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-config-data\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.354460 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-config-data\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.354583 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-kolla-config\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.358397 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-memcached-tls-certs\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.358407 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-combined-ca-bundle\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.373637 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pktm9\" (UniqueName: \"kubernetes.io/projected/131ae7de-5f52-4fd7-8789-d61f8a3a9c46-kube-api-access-pktm9\") pod \"memcached-0\" (UID: \"131ae7de-5f52-4fd7-8789-d61f8a3a9c46\") " pod="openstack/memcached-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.383685 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 15:30:04 crc kubenswrapper[4945]: I1014 15:30:04.627437 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.418474 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.426332 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.438901 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-2dbbm" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.447373 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.485134 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzhmq\" (UniqueName: \"kubernetes.io/projected/42e0b677-3889-4be1-8fbb-8495a2239ade-kube-api-access-wzhmq\") pod \"kube-state-metrics-0\" (UID: \"42e0b677-3889-4be1-8fbb-8495a2239ade\") " pod="openstack/kube-state-metrics-0" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.586793 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzhmq\" (UniqueName: \"kubernetes.io/projected/42e0b677-3889-4be1-8fbb-8495a2239ade-kube-api-access-wzhmq\") pod \"kube-state-metrics-0\" (UID: \"42e0b677-3889-4be1-8fbb-8495a2239ade\") " pod="openstack/kube-state-metrics-0" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.620038 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzhmq\" (UniqueName: \"kubernetes.io/projected/42e0b677-3889-4be1-8fbb-8495a2239ade-kube-api-access-wzhmq\") pod \"kube-state-metrics-0\" (UID: \"42e0b677-3889-4be1-8fbb-8495a2239ade\") " pod="openstack/kube-state-metrics-0" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.662647 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.688061 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a85b7c5c-09cb-413d-a321-18aaa9728a52-config-volume\") pod \"a85b7c5c-09cb-413d-a321-18aaa9728a52\" (UID: \"a85b7c5c-09cb-413d-a321-18aaa9728a52\") " Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.688231 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a85b7c5c-09cb-413d-a321-18aaa9728a52-secret-volume\") pod \"a85b7c5c-09cb-413d-a321-18aaa9728a52\" (UID: \"a85b7c5c-09cb-413d-a321-18aaa9728a52\") " Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.688344 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j997v\" (UniqueName: \"kubernetes.io/projected/a85b7c5c-09cb-413d-a321-18aaa9728a52-kube-api-access-j997v\") pod \"a85b7c5c-09cb-413d-a321-18aaa9728a52\" (UID: \"a85b7c5c-09cb-413d-a321-18aaa9728a52\") " Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.688675 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a85b7c5c-09cb-413d-a321-18aaa9728a52-config-volume" (OuterVolumeSpecName: "config-volume") pod "a85b7c5c-09cb-413d-a321-18aaa9728a52" (UID: "a85b7c5c-09cb-413d-a321-18aaa9728a52"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.693625 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85b7c5c-09cb-413d-a321-18aaa9728a52-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a85b7c5c-09cb-413d-a321-18aaa9728a52" (UID: "a85b7c5c-09cb-413d-a321-18aaa9728a52"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.699113 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a85b7c5c-09cb-413d-a321-18aaa9728a52-kube-api-access-j997v" (OuterVolumeSpecName: "kube-api-access-j997v") pod "a85b7c5c-09cb-413d-a321-18aaa9728a52" (UID: "a85b7c5c-09cb-413d-a321-18aaa9728a52"). InnerVolumeSpecName "kube-api-access-j997v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.789972 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a85b7c5c-09cb-413d-a321-18aaa9728a52-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.790311 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j997v\" (UniqueName: \"kubernetes.io/projected/a85b7c5c-09cb-413d-a321-18aaa9728a52-kube-api-access-j997v\") on node \"crc\" DevicePath \"\"" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.790522 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a85b7c5c-09cb-413d-a321-18aaa9728a52-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 15:30:06 crc kubenswrapper[4945]: I1014 15:30:06.801495 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 15:30:07 crc kubenswrapper[4945]: I1014 15:30:07.268173 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" event={"ID":"a85b7c5c-09cb-413d-a321-18aaa9728a52","Type":"ContainerDied","Data":"64cb2a088ad71d4f8a647bc2cfc4d99f61b4792f798a59a8650e9e56d2741d2b"} Oct 14 15:30:07 crc kubenswrapper[4945]: I1014 15:30:07.268502 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64cb2a088ad71d4f8a647bc2cfc4d99f61b4792f798a59a8650e9e56d2741d2b" Oct 14 15:30:07 crc kubenswrapper[4945]: I1014 15:30:07.268244 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340930-q7k7b" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.256377 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 15:30:10 crc kubenswrapper[4945]: E1014 15:30:10.257020 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85b7c5c-09cb-413d-a321-18aaa9728a52" containerName="collect-profiles" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.257037 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85b7c5c-09cb-413d-a321-18aaa9728a52" containerName="collect-profiles" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.257268 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85b7c5c-09cb-413d-a321-18aaa9728a52" containerName="collect-profiles" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.258229 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.259789 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.260123 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-mrpjx" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.260301 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.260957 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.262571 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.271118 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.341727 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e90e1239-2529-43dd-b429-c140543cb579-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.341767 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e90e1239-2529-43dd-b429-c140543cb579-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.341795 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e90e1239-2529-43dd-b429-c140543cb579-config\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.341942 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e90e1239-2529-43dd-b429-c140543cb579-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.342017 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.342079 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25hsn\" (UniqueName: \"kubernetes.io/projected/e90e1239-2529-43dd-b429-c140543cb579-kube-api-access-25hsn\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.342140 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e90e1239-2529-43dd-b429-c140543cb579-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.342194 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e90e1239-2529-43dd-b429-c140543cb579-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.443283 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.443348 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25hsn\" (UniqueName: \"kubernetes.io/projected/e90e1239-2529-43dd-b429-c140543cb579-kube-api-access-25hsn\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.443380 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e90e1239-2529-43dd-b429-c140543cb579-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.443412 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e90e1239-2529-43dd-b429-c140543cb579-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.443438 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e90e1239-2529-43dd-b429-c140543cb579-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.443461 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e90e1239-2529-43dd-b429-c140543cb579-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.443485 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e90e1239-2529-43dd-b429-c140543cb579-config\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.443535 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e90e1239-2529-43dd-b429-c140543cb579-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.443768 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.452981 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e90e1239-2529-43dd-b429-c140543cb579-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.453243 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e90e1239-2529-43dd-b429-c140543cb579-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.453593 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e90e1239-2529-43dd-b429-c140543cb579-config\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.453628 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e90e1239-2529-43dd-b429-c140543cb579-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.454366 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e90e1239-2529-43dd-b429-c140543cb579-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.456033 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e90e1239-2529-43dd-b429-c140543cb579-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.467113 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25hsn\" (UniqueName: \"kubernetes.io/projected/e90e1239-2529-43dd-b429-c140543cb579-kube-api-access-25hsn\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.468046 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"e90e1239-2529-43dd-b429-c140543cb579\") " pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.578237 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.882809 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6v9cw"] Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.884076 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.887916 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-rjkzf" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.887947 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.888264 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.893643 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-pq8xf"] Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.895935 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.903388 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6v9cw"] Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.915402 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-pq8xf"] Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.948814 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f495edd-59f6-4a56-9356-496cda4ae47a-var-run\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.948922 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d0edbee-18bd-41ac-880b-526477ef54d3-combined-ca-bundle\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.948955 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65w2p\" (UniqueName: \"kubernetes.io/projected/9d0edbee-18bd-41ac-880b-526477ef54d3-kube-api-access-65w2p\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.948974 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2f495edd-59f6-4a56-9356-496cda4ae47a-var-lib\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.949002 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9d0edbee-18bd-41ac-880b-526477ef54d3-var-log-ovn\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.949033 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9d0edbee-18bd-41ac-880b-526477ef54d3-var-run\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.949056 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f495edd-59f6-4a56-9356-496cda4ae47a-scripts\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.949087 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2f495edd-59f6-4a56-9356-496cda4ae47a-etc-ovs\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.949116 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm7ld\" (UniqueName: \"kubernetes.io/projected/2f495edd-59f6-4a56-9356-496cda4ae47a-kube-api-access-bm7ld\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.949147 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9d0edbee-18bd-41ac-880b-526477ef54d3-var-run-ovn\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.949179 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2f495edd-59f6-4a56-9356-496cda4ae47a-var-log\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.949208 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d0edbee-18bd-41ac-880b-526477ef54d3-scripts\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:10 crc kubenswrapper[4945]: I1014 15:30:10.949245 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d0edbee-18bd-41ac-880b-526477ef54d3-ovn-controller-tls-certs\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.050604 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2f495edd-59f6-4a56-9356-496cda4ae47a-var-log\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.050664 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d0edbee-18bd-41ac-880b-526477ef54d3-scripts\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.050708 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d0edbee-18bd-41ac-880b-526477ef54d3-ovn-controller-tls-certs\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.050737 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f495edd-59f6-4a56-9356-496cda4ae47a-var-run\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.050786 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d0edbee-18bd-41ac-880b-526477ef54d3-combined-ca-bundle\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.050811 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65w2p\" (UniqueName: \"kubernetes.io/projected/9d0edbee-18bd-41ac-880b-526477ef54d3-kube-api-access-65w2p\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.050833 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2f495edd-59f6-4a56-9356-496cda4ae47a-var-lib\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.050862 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9d0edbee-18bd-41ac-880b-526477ef54d3-var-log-ovn\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.050917 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9d0edbee-18bd-41ac-880b-526477ef54d3-var-run\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.050937 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f495edd-59f6-4a56-9356-496cda4ae47a-scripts\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.050966 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2f495edd-59f6-4a56-9356-496cda4ae47a-etc-ovs\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.050994 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm7ld\" (UniqueName: \"kubernetes.io/projected/2f495edd-59f6-4a56-9356-496cda4ae47a-kube-api-access-bm7ld\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.051302 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9d0edbee-18bd-41ac-880b-526477ef54d3-var-run-ovn\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.051323 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9d0edbee-18bd-41ac-880b-526477ef54d3-var-run\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.051514 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2f495edd-59f6-4a56-9356-496cda4ae47a-var-lib\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.051195 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2f495edd-59f6-4a56-9356-496cda4ae47a-var-log\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.051547 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2f495edd-59f6-4a56-9356-496cda4ae47a-etc-ovs\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.051583 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f495edd-59f6-4a56-9356-496cda4ae47a-var-run\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.051761 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9d0edbee-18bd-41ac-880b-526477ef54d3-var-log-ovn\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.051778 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9d0edbee-18bd-41ac-880b-526477ef54d3-var-run-ovn\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.052861 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d0edbee-18bd-41ac-880b-526477ef54d3-scripts\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.053190 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f495edd-59f6-4a56-9356-496cda4ae47a-scripts\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.055714 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d0edbee-18bd-41ac-880b-526477ef54d3-ovn-controller-tls-certs\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.067185 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65w2p\" (UniqueName: \"kubernetes.io/projected/9d0edbee-18bd-41ac-880b-526477ef54d3-kube-api-access-65w2p\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.068108 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm7ld\" (UniqueName: \"kubernetes.io/projected/2f495edd-59f6-4a56-9356-496cda4ae47a-kube-api-access-bm7ld\") pod \"ovn-controller-ovs-pq8xf\" (UID: \"2f495edd-59f6-4a56-9356-496cda4ae47a\") " pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.068355 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d0edbee-18bd-41ac-880b-526477ef54d3-combined-ca-bundle\") pod \"ovn-controller-6v9cw\" (UID: \"9d0edbee-18bd-41ac-880b-526477ef54d3\") " pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.235064 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:11 crc kubenswrapper[4945]: I1014 15:30:11.249540 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.737670 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.744450 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.747784 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.748057 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-9r5fr" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.749231 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.749292 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.764281 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.894286 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vng59\" (UniqueName: \"kubernetes.io/projected/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-kube-api-access-vng59\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.894328 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-config\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.894362 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.894419 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.894446 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.894504 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.894524 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.895111 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.996361 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.997347 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.997398 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.997417 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.997474 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.997542 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vng59\" (UniqueName: \"kubernetes.io/projected/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-kube-api-access-vng59\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.997562 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-config\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.997590 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.997792 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:13 crc kubenswrapper[4945]: I1014 15:30:13.998458 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:14 crc kubenswrapper[4945]: I1014 15:30:13.999457 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-config\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:14 crc kubenswrapper[4945]: I1014 15:30:13.999777 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:14 crc kubenswrapper[4945]: I1014 15:30:14.001658 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:14 crc kubenswrapper[4945]: I1014 15:30:14.004233 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:14 crc kubenswrapper[4945]: I1014 15:30:14.017986 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:14 crc kubenswrapper[4945]: I1014 15:30:14.029661 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vng59\" (UniqueName: \"kubernetes.io/projected/2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3-kube-api-access-vng59\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:14 crc kubenswrapper[4945]: I1014 15:30:14.038183 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3\") " pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:14 crc kubenswrapper[4945]: I1014 15:30:14.070125 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 15:30:16 crc kubenswrapper[4945]: I1014 15:30:16.468838 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 15:30:17 crc kubenswrapper[4945]: W1014 15:30:17.098023 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0fa2603_a8a2_449d_88d2_e35975a18ad1.slice/crio-6211f93bb9c816214834584b52f8d14bcd5501368d6867e56f2e8d9968488184 WatchSource:0}: Error finding container 6211f93bb9c816214834584b52f8d14bcd5501368d6867e56f2e8d9968488184: Status 404 returned error can't find the container with id 6211f93bb9c816214834584b52f8d14bcd5501368d6867e56f2e8d9968488184 Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.107141 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 15:30:17 crc kubenswrapper[4945]: E1014 15:30:17.199350 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 14 15:30:17 crc kubenswrapper[4945]: E1014 15:30:17.200402 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mw7xj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-2qkbt_openstack(40d751f7-802d-41ee-9afb-16f33e49f0c0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:30:17 crc kubenswrapper[4945]: E1014 15:30:17.201748 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" podUID="40d751f7-802d-41ee-9afb-16f33e49f0c0" Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.349445 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f0fa2603-a8a2-449d-88d2-e35975a18ad1","Type":"ContainerStarted","Data":"6211f93bb9c816214834584b52f8d14bcd5501368d6867e56f2e8d9968488184"} Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.630910 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.656315 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.688414 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.857745 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40d751f7-802d-41ee-9afb-16f33e49f0c0-dns-svc\") pod \"40d751f7-802d-41ee-9afb-16f33e49f0c0\" (UID: \"40d751f7-802d-41ee-9afb-16f33e49f0c0\") " Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.857961 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mw7xj\" (UniqueName: \"kubernetes.io/projected/40d751f7-802d-41ee-9afb-16f33e49f0c0-kube-api-access-mw7xj\") pod \"40d751f7-802d-41ee-9afb-16f33e49f0c0\" (UID: \"40d751f7-802d-41ee-9afb-16f33e49f0c0\") " Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.857996 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40d751f7-802d-41ee-9afb-16f33e49f0c0-config\") pod \"40d751f7-802d-41ee-9afb-16f33e49f0c0\" (UID: \"40d751f7-802d-41ee-9afb-16f33e49f0c0\") " Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.858369 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40d751f7-802d-41ee-9afb-16f33e49f0c0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "40d751f7-802d-41ee-9afb-16f33e49f0c0" (UID: "40d751f7-802d-41ee-9afb-16f33e49f0c0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.858504 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40d751f7-802d-41ee-9afb-16f33e49f0c0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.858594 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40d751f7-802d-41ee-9afb-16f33e49f0c0-config" (OuterVolumeSpecName: "config") pod "40d751f7-802d-41ee-9afb-16f33e49f0c0" (UID: "40d751f7-802d-41ee-9afb-16f33e49f0c0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.863276 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40d751f7-802d-41ee-9afb-16f33e49f0c0-kube-api-access-mw7xj" (OuterVolumeSpecName: "kube-api-access-mw7xj") pod "40d751f7-802d-41ee-9afb-16f33e49f0c0" (UID: "40d751f7-802d-41ee-9afb-16f33e49f0c0"). InnerVolumeSpecName "kube-api-access-mw7xj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.960517 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mw7xj\" (UniqueName: \"kubernetes.io/projected/40d751f7-802d-41ee-9afb-16f33e49f0c0-kube-api-access-mw7xj\") on node \"crc\" DevicePath \"\"" Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.960561 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40d751f7-802d-41ee-9afb-16f33e49f0c0-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:30:17 crc kubenswrapper[4945]: I1014 15:30:17.979215 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.009528 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.014187 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6v9cw"] Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.018236 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.096038 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 15:30:18 crc kubenswrapper[4945]: E1014 15:30:18.113449 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 14 15:30:18 crc kubenswrapper[4945]: E1014 15:30:18.113601 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5whp9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-pttvf_openstack(9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:30:18 crc kubenswrapper[4945]: E1014 15:30:18.114842 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" podUID="9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48" Oct 14 15:30:18 crc kubenswrapper[4945]: W1014 15:30:18.306344 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod131ae7de_5f52_4fd7_8789_d61f8a3a9c46.slice/crio-2865e20baa325c187ecddd5de529d60424ee7b5b0c4d4f4952048b869e33c3df WatchSource:0}: Error finding container 2865e20baa325c187ecddd5de529d60424ee7b5b0c4d4f4952048b869e33c3df: Status 404 returned error can't find the container with id 2865e20baa325c187ecddd5de529d60424ee7b5b0c4d4f4952048b869e33c3df Oct 14 15:30:18 crc kubenswrapper[4945]: W1014 15:30:18.307384 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d0edbee_18bd_41ac_880b_526477ef54d3.slice/crio-095f36ff882dd394a3b7439b451445849712577146447294cbbabc6d1e58d1fc WatchSource:0}: Error finding container 095f36ff882dd394a3b7439b451445849712577146447294cbbabc6d1e58d1fc: Status 404 returned error can't find the container with id 095f36ff882dd394a3b7439b451445849712577146447294cbbabc6d1e58d1fc Oct 14 15:30:18 crc kubenswrapper[4945]: W1014 15:30:18.313581 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42e0b677_3889_4be1_8fbb_8495a2239ade.slice/crio-06e855ff90a90dc86aacefe05ca369a96d6147838caff960e83c96a7a741d9c8 WatchSource:0}: Error finding container 06e855ff90a90dc86aacefe05ca369a96d6147838caff960e83c96a7a741d9c8: Status 404 returned error can't find the container with id 06e855ff90a90dc86aacefe05ca369a96d6147838caff960e83c96a7a741d9c8 Oct 14 15:30:18 crc kubenswrapper[4945]: W1014 15:30:18.315296 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2195ac97_ef9f_4fd8_b56b_4b96bb1c59f3.slice/crio-6d824957ee4937776e91d1eef4ac97ec0f27fcb085ac4dfdae04142da77c7f19 WatchSource:0}: Error finding container 6d824957ee4937776e91d1eef4ac97ec0f27fcb085ac4dfdae04142da77c7f19: Status 404 returned error can't find the container with id 6d824957ee4937776e91d1eef4ac97ec0f27fcb085ac4dfdae04142da77c7f19 Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.364390 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"afed9c20-1964-44dd-a849-af9fc0f97e86","Type":"ContainerStarted","Data":"11c0acc306860f24890a6722c2cc97f6fd9c5c47a79ff474338e8f252be41326"} Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.365747 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c7915803-bfbc-4150-854d-6039c685a771","Type":"ContainerStarted","Data":"adea549cfc67ed39106c0fc0d2ed36f64d6a4b305ee21ef5341cb432e241c2ed"} Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.366736 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3","Type":"ContainerStarted","Data":"6d824957ee4937776e91d1eef4ac97ec0f27fcb085ac4dfdae04142da77c7f19"} Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.367677 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"42e0b677-3889-4be1-8fbb-8495a2239ade","Type":"ContainerStarted","Data":"06e855ff90a90dc86aacefe05ca369a96d6147838caff960e83c96a7a741d9c8"} Oct 14 15:30:18 crc kubenswrapper[4945]: E1014 15:30:18.368261 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 14 15:30:18 crc kubenswrapper[4945]: E1014 15:30:18.368460 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ppzzv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-pxxrt_openstack(ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.368727 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9642a3f8-eb9b-4f79-8448-602a5f7761d6","Type":"ContainerStarted","Data":"24d4b18d9bb1c46c9fffd46d7196f2130b4779a152df3350e5101c338988cf56"} Oct 14 15:30:18 crc kubenswrapper[4945]: E1014 15:30:18.369773 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" podUID="ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.372399 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"131ae7de-5f52-4fd7-8789-d61f8a3a9c46","Type":"ContainerStarted","Data":"2865e20baa325c187ecddd5de529d60424ee7b5b0c4d4f4952048b869e33c3df"} Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.373827 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6v9cw" event={"ID":"9d0edbee-18bd-41ac-880b-526477ef54d3","Type":"ContainerStarted","Data":"095f36ff882dd394a3b7439b451445849712577146447294cbbabc6d1e58d1fc"} Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.375215 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.375210 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2qkbt" event={"ID":"40d751f7-802d-41ee-9afb-16f33e49f0c0","Type":"ContainerDied","Data":"127ae298f6900314d6ae84df2e70dc3e5187b4e1a9006f0b690a7c95c73a433f"} Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.441802 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2qkbt"] Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.448165 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2qkbt"] Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.634587 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.770291 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48-config\") pod \"9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48\" (UID: \"9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48\") " Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.770393 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5whp9\" (UniqueName: \"kubernetes.io/projected/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48-kube-api-access-5whp9\") pod \"9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48\" (UID: \"9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48\") " Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.771292 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48-config" (OuterVolumeSpecName: "config") pod "9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48" (UID: "9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.775487 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48-kube-api-access-5whp9" (OuterVolumeSpecName: "kube-api-access-5whp9") pod "9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48" (UID: "9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48"). InnerVolumeSpecName "kube-api-access-5whp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.775908 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40d751f7-802d-41ee-9afb-16f33e49f0c0" path="/var/lib/kubelet/pods/40d751f7-802d-41ee-9afb-16f33e49f0c0/volumes" Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.872408 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:30:18 crc kubenswrapper[4945]: I1014 15:30:18.872468 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5whp9\" (UniqueName: \"kubernetes.io/projected/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48-kube-api-access-5whp9\") on node \"crc\" DevicePath \"\"" Oct 14 15:30:19 crc kubenswrapper[4945]: I1014 15:30:19.052751 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-pq8xf"] Oct 14 15:30:19 crc kubenswrapper[4945]: I1014 15:30:19.385645 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" event={"ID":"9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48","Type":"ContainerDied","Data":"f421bb05563e381280a4fb7174f83c4bf11bf5cb1f6ba3546751c1437223048e"} Oct 14 15:30:19 crc kubenswrapper[4945]: I1014 15:30:19.385669 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-pttvf" Oct 14 15:30:19 crc kubenswrapper[4945]: I1014 15:30:19.388087 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pq8xf" event={"ID":"2f495edd-59f6-4a56-9356-496cda4ae47a","Type":"ContainerStarted","Data":"052d72ba37c7331511c3139dd0a53d722c19663c05cc49ab30c66e433d175f2c"} Oct 14 15:30:19 crc kubenswrapper[4945]: I1014 15:30:19.392310 4945 generic.go:334] "Generic (PLEG): container finished" podID="39764f61-839c-4425-9b8b-13e92865b7d2" containerID="1fd88d1780df49018ae4a6358b2aa595893e0f5ee39d30f6da12278e064dda68" exitCode=0 Oct 14 15:30:19 crc kubenswrapper[4945]: I1014 15:30:19.392400 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" event={"ID":"39764f61-839c-4425-9b8b-13e92865b7d2","Type":"ContainerDied","Data":"1fd88d1780df49018ae4a6358b2aa595893e0f5ee39d30f6da12278e064dda68"} Oct 14 15:30:19 crc kubenswrapper[4945]: I1014 15:30:19.426298 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-pttvf"] Oct 14 15:30:19 crc kubenswrapper[4945]: I1014 15:30:19.431652 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-pttvf"] Oct 14 15:30:19 crc kubenswrapper[4945]: I1014 15:30:19.636520 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 15:30:20 crc kubenswrapper[4945]: I1014 15:30:20.403669 4945 generic.go:334] "Generic (PLEG): container finished" podID="ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" containerID="4dcdcfbbfefdde6dbb1b05de5ba9322a0ad0dfe669c233735b92e96edcfb6771" exitCode=0 Oct 14 15:30:20 crc kubenswrapper[4945]: I1014 15:30:20.403790 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" event={"ID":"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c","Type":"ContainerDied","Data":"4dcdcfbbfefdde6dbb1b05de5ba9322a0ad0dfe669c233735b92e96edcfb6771"} Oct 14 15:30:20 crc kubenswrapper[4945]: I1014 15:30:20.407936 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"e90e1239-2529-43dd-b429-c140543cb579","Type":"ContainerStarted","Data":"5234dd93deb5a7ccb95a59c8de5ddbbc97b8778595869d9e26b1ac78e10ea18f"} Oct 14 15:30:20 crc kubenswrapper[4945]: I1014 15:30:20.418218 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" event={"ID":"39764f61-839c-4425-9b8b-13e92865b7d2","Type":"ContainerStarted","Data":"aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff"} Oct 14 15:30:20 crc kubenswrapper[4945]: I1014 15:30:20.418560 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:30:20 crc kubenswrapper[4945]: I1014 15:30:20.456032 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" podStartSLOduration=3.826024081 podStartE2EDuration="21.455995289s" podCreationTimestamp="2025-10-14 15:29:59 +0000 UTC" firstStartedPulling="2025-10-14 15:30:00.802498284 +0000 UTC m=+930.786546652" lastFinishedPulling="2025-10-14 15:30:18.432469492 +0000 UTC m=+948.416517860" observedRunningTime="2025-10-14 15:30:20.442025821 +0000 UTC m=+950.426074269" watchObservedRunningTime="2025-10-14 15:30:20.455995289 +0000 UTC m=+950.440043657" Oct 14 15:30:20 crc kubenswrapper[4945]: I1014 15:30:20.772094 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48" path="/var/lib/kubelet/pods/9ba421df-a3a0-4cdc-bb38-fb2fdbe14d48/volumes" Oct 14 15:30:21 crc kubenswrapper[4945]: I1014 15:30:21.428259 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" event={"ID":"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c","Type":"ContainerStarted","Data":"353dd4f3fda6fb2e1bcc2e2e1ef8aad67b790a9cf97e838ee3a140cd5f6e2a79"} Oct 14 15:30:21 crc kubenswrapper[4945]: I1014 15:30:21.449260 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" podStartSLOduration=-9223372014.405539 podStartE2EDuration="22.449237411s" podCreationTimestamp="2025-10-14 15:29:59 +0000 UTC" firstStartedPulling="2025-10-14 15:30:00.475545348 +0000 UTC m=+930.459593726" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:30:21.446639627 +0000 UTC m=+951.430688005" watchObservedRunningTime="2025-10-14 15:30:21.449237411 +0000 UTC m=+951.433285779" Oct 14 15:30:24 crc kubenswrapper[4945]: I1014 15:30:24.894731 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:30:25 crc kubenswrapper[4945]: I1014 15:30:25.200142 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:30:25 crc kubenswrapper[4945]: I1014 15:30:25.261985 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-pxxrt"] Oct 14 15:30:25 crc kubenswrapper[4945]: I1014 15:30:25.465656 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" podUID="ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" containerName="dnsmasq-dns" containerID="cri-o://353dd4f3fda6fb2e1bcc2e2e1ef8aad67b790a9cf97e838ee3a140cd5f6e2a79" gracePeriod=10 Oct 14 15:30:25 crc kubenswrapper[4945]: I1014 15:30:25.467177 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:30:28 crc kubenswrapper[4945]: I1014 15:30:28.488590 4945 generic.go:334] "Generic (PLEG): container finished" podID="ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" containerID="353dd4f3fda6fb2e1bcc2e2e1ef8aad67b790a9cf97e838ee3a140cd5f6e2a79" exitCode=0 Oct 14 15:30:28 crc kubenswrapper[4945]: I1014 15:30:28.488694 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" event={"ID":"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c","Type":"ContainerDied","Data":"353dd4f3fda6fb2e1bcc2e2e1ef8aad67b790a9cf97e838ee3a140cd5f6e2a79"} Oct 14 15:30:29 crc kubenswrapper[4945]: I1014 15:30:29.894181 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" podUID="ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.97:5353: connect: connection refused" Oct 14 15:30:32 crc kubenswrapper[4945]: I1014 15:30:32.752774 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:30:32 crc kubenswrapper[4945]: E1014 15:30:32.774511 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Oct 14 15:30:32 crc kubenswrapper[4945]: E1014 15:30:32.774798 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-498dl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(f0fa2603-a8a2-449d-88d2-e35975a18ad1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:30:32 crc kubenswrapper[4945]: E1014 15:30:32.776496 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="f0fa2603-a8a2-449d-88d2-e35975a18ad1" Oct 14 15:30:32 crc kubenswrapper[4945]: I1014 15:30:32.928276 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppzzv\" (UniqueName: \"kubernetes.io/projected/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-kube-api-access-ppzzv\") pod \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\" (UID: \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\") " Oct 14 15:30:32 crc kubenswrapper[4945]: I1014 15:30:32.928471 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-dns-svc\") pod \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\" (UID: \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\") " Oct 14 15:30:32 crc kubenswrapper[4945]: I1014 15:30:32.928509 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-config\") pod \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\" (UID: \"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c\") " Oct 14 15:30:32 crc kubenswrapper[4945]: I1014 15:30:32.934076 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-kube-api-access-ppzzv" (OuterVolumeSpecName: "kube-api-access-ppzzv") pod "ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" (UID: "ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c"). InnerVolumeSpecName "kube-api-access-ppzzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:30:32 crc kubenswrapper[4945]: I1014 15:30:32.967427 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" (UID: "ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:30:32 crc kubenswrapper[4945]: I1014 15:30:32.971493 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-config" (OuterVolumeSpecName: "config") pod "ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" (UID: "ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:30:33 crc kubenswrapper[4945]: I1014 15:30:33.030013 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppzzv\" (UniqueName: \"kubernetes.io/projected/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-kube-api-access-ppzzv\") on node \"crc\" DevicePath \"\"" Oct 14 15:30:33 crc kubenswrapper[4945]: I1014 15:30:33.030090 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:30:33 crc kubenswrapper[4945]: I1014 15:30:33.030124 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:30:33 crc kubenswrapper[4945]: I1014 15:30:33.531085 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" Oct 14 15:30:33 crc kubenswrapper[4945]: I1014 15:30:33.531086 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-pxxrt" event={"ID":"ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c","Type":"ContainerDied","Data":"b8a4b4c3ab87d54e2aea7e733ddc10e4b9602d8bd5b8130527b60d655b4a71a7"} Oct 14 15:30:33 crc kubenswrapper[4945]: I1014 15:30:33.531302 4945 scope.go:117] "RemoveContainer" containerID="353dd4f3fda6fb2e1bcc2e2e1ef8aad67b790a9cf97e838ee3a140cd5f6e2a79" Oct 14 15:30:33 crc kubenswrapper[4945]: I1014 15:30:33.570703 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-pxxrt"] Oct 14 15:30:33 crc kubenswrapper[4945]: I1014 15:30:33.575280 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-pxxrt"] Oct 14 15:30:33 crc kubenswrapper[4945]: E1014 15:30:33.632250 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="f0fa2603-a8a2-449d-88d2-e35975a18ad1" Oct 14 15:30:33 crc kubenswrapper[4945]: E1014 15:30:33.845768 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Oct 14 15:30:33 crc kubenswrapper[4945]: E1014 15:30:33.846032 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bd9fl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(afed9c20-1964-44dd-a849-af9fc0f97e86): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:30:33 crc kubenswrapper[4945]: E1014 15:30:33.847269 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="afed9c20-1964-44dd-a849-af9fc0f97e86" Oct 14 15:30:34 crc kubenswrapper[4945]: E1014 15:30:34.539065 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="afed9c20-1964-44dd-a849-af9fc0f97e86" Oct 14 15:30:34 crc kubenswrapper[4945]: I1014 15:30:34.771504 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" path="/var/lib/kubelet/pods/ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c/volumes" Oct 14 15:30:37 crc kubenswrapper[4945]: I1014 15:30:37.642657 4945 scope.go:117] "RemoveContainer" containerID="4dcdcfbbfefdde6dbb1b05de5ba9322a0ad0dfe669c233735b92e96edcfb6771" Oct 14 15:30:40 crc kubenswrapper[4945]: I1014 15:30:40.588243 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pq8xf" event={"ID":"2f495edd-59f6-4a56-9356-496cda4ae47a","Type":"ContainerStarted","Data":"7f429abf99a75991d2b46f2beb34d80b48598517878e07d09ae79e4ec20736e7"} Oct 14 15:30:40 crc kubenswrapper[4945]: I1014 15:30:40.591520 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"131ae7de-5f52-4fd7-8789-d61f8a3a9c46","Type":"ContainerStarted","Data":"e688ad68f27c82832f8d103f2762a270471412baa485592907e240e94c612edd"} Oct 14 15:30:40 crc kubenswrapper[4945]: I1014 15:30:40.591743 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 14 15:30:40 crc kubenswrapper[4945]: I1014 15:30:40.627484 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=17.053408199 podStartE2EDuration="36.62746639s" podCreationTimestamp="2025-10-14 15:30:04 +0000 UTC" firstStartedPulling="2025-10-14 15:30:18.309071272 +0000 UTC m=+948.293119640" lastFinishedPulling="2025-10-14 15:30:37.883129463 +0000 UTC m=+967.867177831" observedRunningTime="2025-10-14 15:30:40.627218083 +0000 UTC m=+970.611266461" watchObservedRunningTime="2025-10-14 15:30:40.62746639 +0000 UTC m=+970.611514758" Oct 14 15:30:41 crc kubenswrapper[4945]: I1014 15:30:41.598979 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"e90e1239-2529-43dd-b429-c140543cb579","Type":"ContainerStarted","Data":"9fc78266992d6124e6714667ed4520004405b2d6ebfaf586294c1775c2858d29"} Oct 14 15:30:41 crc kubenswrapper[4945]: I1014 15:30:41.600266 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6v9cw" event={"ID":"9d0edbee-18bd-41ac-880b-526477ef54d3","Type":"ContainerStarted","Data":"6c3dbfdeaca4f22111cfc8816e358b9cfc9b13bc3292e9463f374cfb1fbae5f0"} Oct 14 15:30:41 crc kubenswrapper[4945]: I1014 15:30:41.600350 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-6v9cw" Oct 14 15:30:41 crc kubenswrapper[4945]: I1014 15:30:41.602438 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c7915803-bfbc-4150-854d-6039c685a771","Type":"ContainerStarted","Data":"9c60658f6099b1c229bfd7f9e639bfeb03e6915d14caaeb7e9f7520dc1e46c01"} Oct 14 15:30:41 crc kubenswrapper[4945]: I1014 15:30:41.604154 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3","Type":"ContainerStarted","Data":"d631aec48147083cc52462a0cb018883f3878a08dbaa580a643aacbda50f9248"} Oct 14 15:30:41 crc kubenswrapper[4945]: I1014 15:30:41.605319 4945 generic.go:334] "Generic (PLEG): container finished" podID="2f495edd-59f6-4a56-9356-496cda4ae47a" containerID="7f429abf99a75991d2b46f2beb34d80b48598517878e07d09ae79e4ec20736e7" exitCode=0 Oct 14 15:30:41 crc kubenswrapper[4945]: I1014 15:30:41.606362 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pq8xf" event={"ID":"2f495edd-59f6-4a56-9356-496cda4ae47a","Type":"ContainerDied","Data":"7f429abf99a75991d2b46f2beb34d80b48598517878e07d09ae79e4ec20736e7"} Oct 14 15:30:41 crc kubenswrapper[4945]: I1014 15:30:41.625288 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6v9cw" podStartSLOduration=11.272277393 podStartE2EDuration="31.625268271s" podCreationTimestamp="2025-10-14 15:30:10 +0000 UTC" firstStartedPulling="2025-10-14 15:30:18.309542385 +0000 UTC m=+948.293590753" lastFinishedPulling="2025-10-14 15:30:38.662533273 +0000 UTC m=+968.646581631" observedRunningTime="2025-10-14 15:30:41.618488928 +0000 UTC m=+971.602537316" watchObservedRunningTime="2025-10-14 15:30:41.625268271 +0000 UTC m=+971.609316639" Oct 14 15:30:42 crc kubenswrapper[4945]: I1014 15:30:42.628937 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pq8xf" event={"ID":"2f495edd-59f6-4a56-9356-496cda4ae47a","Type":"ContainerStarted","Data":"667cd7293e8642ff62c53f504168311d5e729c4665b3e73b9f4c3764ee5bd2a3"} Oct 14 15:30:42 crc kubenswrapper[4945]: I1014 15:30:42.631092 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"42e0b677-3889-4be1-8fbb-8495a2239ade","Type":"ContainerStarted","Data":"350cba45b4f29b9a5afac152ad30712ee1e1881184f779bb06e575a83dd6e046"} Oct 14 15:30:42 crc kubenswrapper[4945]: I1014 15:30:42.631862 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 14 15:30:42 crc kubenswrapper[4945]: I1014 15:30:42.638085 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9642a3f8-eb9b-4f79-8448-602a5f7761d6","Type":"ContainerStarted","Data":"421c62871dd88c2e8493610a2bf3454c6bdaab0cc2d9530739b1bc7a8d4d5e01"} Oct 14 15:30:42 crc kubenswrapper[4945]: I1014 15:30:42.658151 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=13.927516125 podStartE2EDuration="36.658128431s" podCreationTimestamp="2025-10-14 15:30:06 +0000 UTC" firstStartedPulling="2025-10-14 15:30:18.428472358 +0000 UTC m=+948.412520726" lastFinishedPulling="2025-10-14 15:30:41.159084664 +0000 UTC m=+971.143133032" observedRunningTime="2025-10-14 15:30:42.657514784 +0000 UTC m=+972.641563152" watchObservedRunningTime="2025-10-14 15:30:42.658128431 +0000 UTC m=+972.642176799" Oct 14 15:30:44 crc kubenswrapper[4945]: I1014 15:30:44.659974 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pq8xf" event={"ID":"2f495edd-59f6-4a56-9356-496cda4ae47a","Type":"ContainerStarted","Data":"388b9d60a67bceed6337675e7ebedb2cefb47f3a919c3f4f12fcd600e25ac18f"} Oct 14 15:30:45 crc kubenswrapper[4945]: I1014 15:30:45.667688 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:45 crc kubenswrapper[4945]: I1014 15:30:45.668769 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:30:45 crc kubenswrapper[4945]: I1014 15:30:45.692228 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-pq8xf" podStartSLOduration=16.874914779 podStartE2EDuration="35.692200633s" podCreationTimestamp="2025-10-14 15:30:10 +0000 UTC" firstStartedPulling="2025-10-14 15:30:19.065365165 +0000 UTC m=+949.049413533" lastFinishedPulling="2025-10-14 15:30:37.882651019 +0000 UTC m=+967.866699387" observedRunningTime="2025-10-14 15:30:45.687730006 +0000 UTC m=+975.671778414" watchObservedRunningTime="2025-10-14 15:30:45.692200633 +0000 UTC m=+975.676249041" Oct 14 15:30:46 crc kubenswrapper[4945]: I1014 15:30:46.805844 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 14 15:30:49 crc kubenswrapper[4945]: I1014 15:30:49.628433 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.533838 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-krh6j"] Oct 14 15:30:54 crc kubenswrapper[4945]: E1014 15:30:54.534733 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" containerName="dnsmasq-dns" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.534753 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" containerName="dnsmasq-dns" Oct 14 15:30:54 crc kubenswrapper[4945]: E1014 15:30:54.534780 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" containerName="init" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.534788 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" containerName="init" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.535027 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea8fbc4f-f6ee-405b-9cf3-e88cce3f5b8c" containerName="dnsmasq-dns" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.535560 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.539719 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.542648 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-krh6j"] Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.608730 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-combined-ca-bundle\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.608806 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.608850 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-config\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.608890 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-ovn-rundir\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.608924 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b44jx\" (UniqueName: \"kubernetes.io/projected/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-kube-api-access-b44jx\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.608989 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-ovs-rundir\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.701791 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-fq22d"] Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.703129 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.705646 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.710425 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-combined-ca-bundle\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.710499 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.710550 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-config\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.710573 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-ovn-rundir\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.710606 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b44jx\" (UniqueName: \"kubernetes.io/projected/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-kube-api-access-b44jx\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.710635 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-ovs-rundir\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.710913 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-ovn-rundir\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.710964 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-ovs-rundir\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.711309 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-config\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.718213 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-combined-ca-bundle\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.720823 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.723484 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-fq22d"] Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.744434 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b44jx\" (UniqueName: \"kubernetes.io/projected/80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4-kube-api-access-b44jx\") pod \"ovn-controller-metrics-krh6j\" (UID: \"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4\") " pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.811752 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-fq22d\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.811896 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lws6g\" (UniqueName: \"kubernetes.io/projected/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-kube-api-access-lws6g\") pod \"dnsmasq-dns-5bf47b49b7-fq22d\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.811971 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-fq22d\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.812016 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-config\") pod \"dnsmasq-dns-5bf47b49b7-fq22d\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.876617 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-krh6j" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.914116 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lws6g\" (UniqueName: \"kubernetes.io/projected/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-kube-api-access-lws6g\") pod \"dnsmasq-dns-5bf47b49b7-fq22d\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.914735 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-fq22d\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.914909 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-config\") pod \"dnsmasq-dns-5bf47b49b7-fq22d\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.915576 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-fq22d\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.916032 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-config\") pod \"dnsmasq-dns-5bf47b49b7-fq22d\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.916217 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-fq22d\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.917395 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-fq22d\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:54 crc kubenswrapper[4945]: I1014 15:30:54.967230 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lws6g\" (UniqueName: \"kubernetes.io/projected/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-kube-api-access-lws6g\") pod \"dnsmasq-dns-5bf47b49b7-fq22d\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.033173 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-fq22d"] Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.034132 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.134276 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-zpn2l"] Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.135794 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.137808 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.145664 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-zpn2l"] Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.220430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-dns-svc\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.220517 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-config\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.220704 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.220788 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.220932 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bndq2\" (UniqueName: \"kubernetes.io/projected/5e19d768-e359-48e1-9456-ede0e7276568-kube-api-access-bndq2\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.322367 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-dns-svc\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.322449 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-config\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.322483 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.322508 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.322551 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bndq2\" (UniqueName: \"kubernetes.io/projected/5e19d768-e359-48e1-9456-ede0e7276568-kube-api-access-bndq2\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.408917 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-dns-svc\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.408986 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.409111 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.409145 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-config\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.412778 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bndq2\" (UniqueName: \"kubernetes.io/projected/5e19d768-e359-48e1-9456-ede0e7276568-kube-api-access-bndq2\") pod \"dnsmasq-dns-8554648995-zpn2l\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:55 crc kubenswrapper[4945]: I1014 15:30:55.509841 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:30:56 crc kubenswrapper[4945]: I1014 15:30:56.843136 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-zpn2l"] Oct 14 15:30:56 crc kubenswrapper[4945]: I1014 15:30:56.876012 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-psqrj"] Oct 14 15:30:56 crc kubenswrapper[4945]: I1014 15:30:56.877439 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:56 crc kubenswrapper[4945]: I1014 15:30:56.895491 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-psqrj"] Oct 14 15:30:56 crc kubenswrapper[4945]: I1014 15:30:56.952687 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:56 crc kubenswrapper[4945]: I1014 15:30:56.952741 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh5vj\" (UniqueName: \"kubernetes.io/projected/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-kube-api-access-mh5vj\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:56 crc kubenswrapper[4945]: I1014 15:30:56.952763 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:56 crc kubenswrapper[4945]: I1014 15:30:56.952811 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-config\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:56 crc kubenswrapper[4945]: I1014 15:30:56.952838 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.054614 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.054670 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh5vj\" (UniqueName: \"kubernetes.io/projected/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-kube-api-access-mh5vj\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.054692 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.054742 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-config\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.054771 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.055768 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.056035 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.057183 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-config\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.057971 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.079111 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh5vj\" (UniqueName: \"kubernetes.io/projected/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-kube-api-access-mh5vj\") pod \"dnsmasq-dns-b8fbc5445-psqrj\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.209675 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.965661 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.983018 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.990622 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.990630 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.990970 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-jgrfl" Oct 14 15:30:57 crc kubenswrapper[4945]: I1014 15:30:57.991223 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.015860 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.069130 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.069186 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhwnk\" (UniqueName: \"kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-kube-api-access-rhwnk\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.069234 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.069272 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/eb8ab92b-e774-40b3-becd-458958582bef-lock\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.069400 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/eb8ab92b-e774-40b3-becd-458958582bef-cache\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.106799 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-zpn2l"] Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.171563 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-krh6j"] Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.172296 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/eb8ab92b-e774-40b3-becd-458958582bef-cache\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.172355 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.172382 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhwnk\" (UniqueName: \"kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-kube-api-access-rhwnk\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.172409 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.172443 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/eb8ab92b-e774-40b3-becd-458958582bef-lock\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.173448 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/eb8ab92b-e774-40b3-becd-458958582bef-lock\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: E1014 15:30:58.173565 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 15:30:58 crc kubenswrapper[4945]: E1014 15:30:58.173582 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 15:30:58 crc kubenswrapper[4945]: E1014 15:30:58.173627 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift podName:eb8ab92b-e774-40b3-becd-458958582bef nodeName:}" failed. No retries permitted until 2025-10-14 15:30:58.673610497 +0000 UTC m=+988.657658865 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift") pod "swift-storage-0" (UID: "eb8ab92b-e774-40b3-becd-458958582bef") : configmap "swift-ring-files" not found Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.173971 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/eb8ab92b-e774-40b3-becd-458958582bef-cache\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.174037 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.200755 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-fq22d"] Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.201048 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhwnk\" (UniqueName: \"kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-kube-api-access-rhwnk\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.205328 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.296892 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-psqrj"] Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.682800 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:58 crc kubenswrapper[4945]: E1014 15:30:58.683042 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 15:30:58 crc kubenswrapper[4945]: E1014 15:30:58.683063 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 15:30:58 crc kubenswrapper[4945]: E1014 15:30:58.683104 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift podName:eb8ab92b-e774-40b3-becd-458958582bef nodeName:}" failed. No retries permitted until 2025-10-14 15:30:59.683090929 +0000 UTC m=+989.667139297 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift") pod "swift-storage-0" (UID: "eb8ab92b-e774-40b3-becd-458958582bef") : configmap "swift-ring-files" not found Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.788763 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"afed9c20-1964-44dd-a849-af9fc0f97e86","Type":"ContainerStarted","Data":"1e1fdce4a59e17ecffd07fc59b3af2de23448bc2dd1a7c08321b85fc31454ded"} Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.793330 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" event={"ID":"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de","Type":"ContainerStarted","Data":"5772dc710ccf11ab15382d20d1c400fcea030c0fd2d0af5369ae1cc601614b15"} Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.793366 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" event={"ID":"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de","Type":"ContainerStarted","Data":"86fc4598270ad28039c1740786b441a4a4e134e8c8596dcaf8bac4c287386ba9"} Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.795898 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f0fa2603-a8a2-449d-88d2-e35975a18ad1","Type":"ContainerStarted","Data":"bf1515b09e7a1e03f11b7561d82030ffe6d87573a28520ac76393663921ce742"} Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.798402 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-zpn2l" event={"ID":"5e19d768-e359-48e1-9456-ede0e7276568","Type":"ContainerStarted","Data":"18dbff8600b36b0f68e390b8bb4498e92e04e284390e724241db14261215e5bf"} Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.798424 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-zpn2l" event={"ID":"5e19d768-e359-48e1-9456-ede0e7276568","Type":"ContainerStarted","Data":"e59b373c473502a2fb2e4e6bb77d6e22806c875639d8d0fc56fc72db6cb35ebc"} Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.802716 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-krh6j" event={"ID":"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4","Type":"ContainerStarted","Data":"b33fb45f8962c9070fdb08af9e264fbb0327bdc0fdd56638ba98c1b69c77d0f0"} Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.805990 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" event={"ID":"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc","Type":"ContainerStarted","Data":"cbcdd1b00aace6e007e092ce8be56bb62ec218c99b74c47ceb71af5b011cbab5"} Oct 14 15:30:58 crc kubenswrapper[4945]: I1014 15:30:58.806053 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" event={"ID":"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc","Type":"ContainerStarted","Data":"c15cd82c0982d63df1d6f0326d62087c352b914ed3d8782475eb41fe6b18038f"} Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.703250 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:30:59 crc kubenswrapper[4945]: E1014 15:30:59.703678 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 15:30:59 crc kubenswrapper[4945]: E1014 15:30:59.703766 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 15:30:59 crc kubenswrapper[4945]: E1014 15:30:59.704001 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift podName:eb8ab92b-e774-40b3-becd-458958582bef nodeName:}" failed. No retries permitted until 2025-10-14 15:31:01.703848634 +0000 UTC m=+991.687897002 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift") pod "swift-storage-0" (UID: "eb8ab92b-e774-40b3-becd-458958582bef") : configmap "swift-ring-files" not found Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.814517 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-krh6j" event={"ID":"80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4","Type":"ContainerStarted","Data":"52287c92c3ca66d36b951b4ce8a1ed1c35ea14d62e999113ee7415d34c013b14"} Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.816340 4945 generic.go:334] "Generic (PLEG): container finished" podID="a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" containerID="cbcdd1b00aace6e007e092ce8be56bb62ec218c99b74c47ceb71af5b011cbab5" exitCode=0 Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.816401 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" event={"ID":"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc","Type":"ContainerDied","Data":"cbcdd1b00aace6e007e092ce8be56bb62ec218c99b74c47ceb71af5b011cbab5"} Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.818196 4945 generic.go:334] "Generic (PLEG): container finished" podID="72d7c6ac-fb71-417c-9c30-8e8eee4ba9de" containerID="5772dc710ccf11ab15382d20d1c400fcea030c0fd2d0af5369ae1cc601614b15" exitCode=0 Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.818252 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" event={"ID":"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de","Type":"ContainerDied","Data":"5772dc710ccf11ab15382d20d1c400fcea030c0fd2d0af5369ae1cc601614b15"} Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.820295 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3","Type":"ContainerStarted","Data":"f1a1652a2b4c267e82da8fe5d605616d11e017485e85c72140e2abf8740ca64d"} Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.822228 4945 generic.go:334] "Generic (PLEG): container finished" podID="5e19d768-e359-48e1-9456-ede0e7276568" containerID="18dbff8600b36b0f68e390b8bb4498e92e04e284390e724241db14261215e5bf" exitCode=0 Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.822274 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-zpn2l" event={"ID":"5e19d768-e359-48e1-9456-ede0e7276568","Type":"ContainerDied","Data":"18dbff8600b36b0f68e390b8bb4498e92e04e284390e724241db14261215e5bf"} Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.824837 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"e90e1239-2529-43dd-b429-c140543cb579","Type":"ContainerStarted","Data":"df7d5ff05f622d7e6b7021ba36e36baa61bbf5462f14e08ebac3917acc80231d"} Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.844677 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-krh6j" podStartSLOduration=5.84465643 podStartE2EDuration="5.84465643s" podCreationTimestamp="2025-10-14 15:30:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:30:59.832623597 +0000 UTC m=+989.816671965" watchObservedRunningTime="2025-10-14 15:30:59.84465643 +0000 UTC m=+989.828704798" Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.882033 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=7.761367266 podStartE2EDuration="47.882013056s" podCreationTimestamp="2025-10-14 15:30:12 +0000 UTC" firstStartedPulling="2025-10-14 15:30:18.42854025 +0000 UTC m=+948.412588618" lastFinishedPulling="2025-10-14 15:30:58.54918604 +0000 UTC m=+988.533234408" observedRunningTime="2025-10-14 15:30:59.859245006 +0000 UTC m=+989.843293374" watchObservedRunningTime="2025-10-14 15:30:59.882013056 +0000 UTC m=+989.866061424" Oct 14 15:30:59 crc kubenswrapper[4945]: I1014 15:30:59.885121 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=11.940762887 podStartE2EDuration="50.885113144s" podCreationTimestamp="2025-10-14 15:30:09 +0000 UTC" firstStartedPulling="2025-10-14 15:30:19.650582907 +0000 UTC m=+949.634631285" lastFinishedPulling="2025-10-14 15:30:58.594933174 +0000 UTC m=+988.578981542" observedRunningTime="2025-10-14 15:30:59.878869086 +0000 UTC m=+989.862917554" watchObservedRunningTime="2025-10-14 15:30:59.885113144 +0000 UTC m=+989.869161512" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.334209 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.420328 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.524533 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bndq2\" (UniqueName: \"kubernetes.io/projected/5e19d768-e359-48e1-9456-ede0e7276568-kube-api-access-bndq2\") pod \"5e19d768-e359-48e1-9456-ede0e7276568\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.524596 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-dns-svc\") pod \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.524632 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-ovsdbserver-nb\") pod \"5e19d768-e359-48e1-9456-ede0e7276568\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.524733 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-ovsdbserver-nb\") pod \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.524757 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-ovsdbserver-sb\") pod \"5e19d768-e359-48e1-9456-ede0e7276568\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.524809 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-config\") pod \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.524950 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-config\") pod \"5e19d768-e359-48e1-9456-ede0e7276568\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.524993 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-dns-svc\") pod \"5e19d768-e359-48e1-9456-ede0e7276568\" (UID: \"5e19d768-e359-48e1-9456-ede0e7276568\") " Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.525030 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lws6g\" (UniqueName: \"kubernetes.io/projected/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-kube-api-access-lws6g\") pod \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\" (UID: \"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de\") " Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.530013 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e19d768-e359-48e1-9456-ede0e7276568-kube-api-access-bndq2" (OuterVolumeSpecName: "kube-api-access-bndq2") pod "5e19d768-e359-48e1-9456-ede0e7276568" (UID: "5e19d768-e359-48e1-9456-ede0e7276568"). InnerVolumeSpecName "kube-api-access-bndq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.530504 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-kube-api-access-lws6g" (OuterVolumeSpecName: "kube-api-access-lws6g") pod "72d7c6ac-fb71-417c-9c30-8e8eee4ba9de" (UID: "72d7c6ac-fb71-417c-9c30-8e8eee4ba9de"). InnerVolumeSpecName "kube-api-access-lws6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.545738 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-config" (OuterVolumeSpecName: "config") pod "72d7c6ac-fb71-417c-9c30-8e8eee4ba9de" (UID: "72d7c6ac-fb71-417c-9c30-8e8eee4ba9de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.548865 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5e19d768-e359-48e1-9456-ede0e7276568" (UID: "5e19d768-e359-48e1-9456-ede0e7276568"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.550633 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5e19d768-e359-48e1-9456-ede0e7276568" (UID: "5e19d768-e359-48e1-9456-ede0e7276568"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.552238 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "72d7c6ac-fb71-417c-9c30-8e8eee4ba9de" (UID: "72d7c6ac-fb71-417c-9c30-8e8eee4ba9de"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.552774 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5e19d768-e359-48e1-9456-ede0e7276568" (UID: "5e19d768-e359-48e1-9456-ede0e7276568"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.557799 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-config" (OuterVolumeSpecName: "config") pod "5e19d768-e359-48e1-9456-ede0e7276568" (UID: "5e19d768-e359-48e1-9456-ede0e7276568"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.566406 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "72d7c6ac-fb71-417c-9c30-8e8eee4ba9de" (UID: "72d7c6ac-fb71-417c-9c30-8e8eee4ba9de"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.578735 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.627032 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.627072 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.627084 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.627096 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.627105 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.627115 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lws6g\" (UniqueName: \"kubernetes.io/projected/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-kube-api-access-lws6g\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.627128 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bndq2\" (UniqueName: \"kubernetes.io/projected/5e19d768-e359-48e1-9456-ede0e7276568-kube-api-access-bndq2\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.627139 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.627148 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e19d768-e359-48e1-9456-ede0e7276568-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.837615 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" event={"ID":"72d7c6ac-fb71-417c-9c30-8e8eee4ba9de","Type":"ContainerDied","Data":"86fc4598270ad28039c1740786b441a4a4e134e8c8596dcaf8bac4c287386ba9"} Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.837714 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-fq22d" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.837740 4945 scope.go:117] "RemoveContainer" containerID="5772dc710ccf11ab15382d20d1c400fcea030c0fd2d0af5369ae1cc601614b15" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.842594 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-zpn2l" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.842595 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-zpn2l" event={"ID":"5e19d768-e359-48e1-9456-ede0e7276568","Type":"ContainerDied","Data":"e59b373c473502a2fb2e4e6bb77d6e22806c875639d8d0fc56fc72db6cb35ebc"} Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.845060 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" event={"ID":"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc","Type":"ContainerStarted","Data":"d37ceab0048ce9492a641679050a5ad77feb1dcceca706ec58186f08af45e444"} Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.846389 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.866167 4945 scope.go:117] "RemoveContainer" containerID="18dbff8600b36b0f68e390b8bb4498e92e04e284390e724241db14261215e5bf" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.867551 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" podStartSLOduration=4.867532176 podStartE2EDuration="4.867532176s" podCreationTimestamp="2025-10-14 15:30:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:31:00.863254834 +0000 UTC m=+990.847303202" watchObservedRunningTime="2025-10-14 15:31:00.867532176 +0000 UTC m=+990.851580544" Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.899757 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-fq22d"] Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.905463 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-fq22d"] Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.930992 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-zpn2l"] Oct 14 15:31:00 crc kubenswrapper[4945]: I1014 15:31:00.935747 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-zpn2l"] Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.578753 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.618931 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.745554 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:31:01 crc kubenswrapper[4945]: E1014 15:31:01.745726 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 15:31:01 crc kubenswrapper[4945]: E1014 15:31:01.745753 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 15:31:01 crc kubenswrapper[4945]: E1014 15:31:01.745808 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift podName:eb8ab92b-e774-40b3-becd-458958582bef nodeName:}" failed. No retries permitted until 2025-10-14 15:31:05.745791068 +0000 UTC m=+995.729839436 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift") pod "swift-storage-0" (UID: "eb8ab92b-e774-40b3-becd-458958582bef") : configmap "swift-ring-files" not found Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.890699 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.956211 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-h5b5w"] Oct 14 15:31:01 crc kubenswrapper[4945]: E1014 15:31:01.956600 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d7c6ac-fb71-417c-9c30-8e8eee4ba9de" containerName="init" Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.956620 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d7c6ac-fb71-417c-9c30-8e8eee4ba9de" containerName="init" Oct 14 15:31:01 crc kubenswrapper[4945]: E1014 15:31:01.956635 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e19d768-e359-48e1-9456-ede0e7276568" containerName="init" Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.956641 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e19d768-e359-48e1-9456-ede0e7276568" containerName="init" Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.956818 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d7c6ac-fb71-417c-9c30-8e8eee4ba9de" containerName="init" Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.956832 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e19d768-e359-48e1-9456-ede0e7276568" containerName="init" Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.957349 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.962331 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.962431 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.962637 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 14 15:31:01 crc kubenswrapper[4945]: I1014 15:31:01.979749 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-h5b5w"] Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.070750 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.109267 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.151974 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-etc-swift\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.152263 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-scripts\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.152346 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-combined-ca-bundle\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.152463 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-ring-data-devices\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.152604 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-dispersionconf\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.152695 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-swiftconf\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.152763 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsjwh\" (UniqueName: \"kubernetes.io/projected/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-kube-api-access-fsjwh\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.255011 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-etc-swift\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.255151 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-scripts\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.255185 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-combined-ca-bundle\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.255229 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-ring-data-devices\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.255245 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-dispersionconf\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.255271 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-swiftconf\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.255286 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsjwh\" (UniqueName: \"kubernetes.io/projected/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-kube-api-access-fsjwh\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.256186 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-etc-swift\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.256433 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-ring-data-devices\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.256459 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-scripts\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.259852 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-dispersionconf\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.260753 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-combined-ca-bundle\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.265762 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-swiftconf\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.280698 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsjwh\" (UniqueName: \"kubernetes.io/projected/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-kube-api-access-fsjwh\") pod \"swift-ring-rebalance-h5b5w\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.576027 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.775043 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e19d768-e359-48e1-9456-ede0e7276568" path="/var/lib/kubelet/pods/5e19d768-e359-48e1-9456-ede0e7276568/volumes" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.775850 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72d7c6ac-fb71-417c-9c30-8e8eee4ba9de" path="/var/lib/kubelet/pods/72d7c6ac-fb71-417c-9c30-8e8eee4ba9de/volumes" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.860981 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 14 15:31:02 crc kubenswrapper[4945]: I1014 15:31:02.900711 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.034434 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.035816 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.039432 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-p7fv8" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.039478 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.039432 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.039643 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.054058 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-h5b5w"] Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.059762 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.175998 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.176063 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.176089 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.176129 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.176165 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-scripts\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.176207 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp9px\" (UniqueName: \"kubernetes.io/projected/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-kube-api-access-qp9px\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.176226 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-config\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.277858 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-scripts\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.277957 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp9px\" (UniqueName: \"kubernetes.io/projected/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-kube-api-access-qp9px\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.277987 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-config\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.278041 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.278073 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.278099 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.278136 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.278517 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.278856 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-config\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.279612 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-scripts\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.285744 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.292570 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.292766 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.299600 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp9px\" (UniqueName: \"kubernetes.io/projected/3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31-kube-api-access-qp9px\") pod \"ovn-northd-0\" (UID: \"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31\") " pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.387381 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.828284 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 15:31:03 crc kubenswrapper[4945]: W1014 15:31:03.833969 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d7482e7_31e7_4d3f_ae1c_84b8ce31ed31.slice/crio-9aa796afecf605c49abe9ee3c0681f2501f1119f8983f9a864b63bf965a98318 WatchSource:0}: Error finding container 9aa796afecf605c49abe9ee3c0681f2501f1119f8983f9a864b63bf965a98318: Status 404 returned error can't find the container with id 9aa796afecf605c49abe9ee3c0681f2501f1119f8983f9a864b63bf965a98318 Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.868146 4945 generic.go:334] "Generic (PLEG): container finished" podID="afed9c20-1964-44dd-a849-af9fc0f97e86" containerID="1e1fdce4a59e17ecffd07fc59b3af2de23448bc2dd1a7c08321b85fc31454ded" exitCode=0 Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.868187 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"afed9c20-1964-44dd-a849-af9fc0f97e86","Type":"ContainerDied","Data":"1e1fdce4a59e17ecffd07fc59b3af2de23448bc2dd1a7c08321b85fc31454ded"} Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.869973 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31","Type":"ContainerStarted","Data":"9aa796afecf605c49abe9ee3c0681f2501f1119f8983f9a864b63bf965a98318"} Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.872046 4945 generic.go:334] "Generic (PLEG): container finished" podID="f0fa2603-a8a2-449d-88d2-e35975a18ad1" containerID="bf1515b09e7a1e03f11b7561d82030ffe6d87573a28520ac76393663921ce742" exitCode=0 Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.872148 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f0fa2603-a8a2-449d-88d2-e35975a18ad1","Type":"ContainerDied","Data":"bf1515b09e7a1e03f11b7561d82030ffe6d87573a28520ac76393663921ce742"} Oct 14 15:31:03 crc kubenswrapper[4945]: I1014 15:31:03.873299 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-h5b5w" event={"ID":"787ea529-5875-4e6c-8a3e-5b4fbf019fa8","Type":"ContainerStarted","Data":"b3fe26bf6d5bdba7358eb3a61a7037e1712ab5b4c8e45524920de8fb1a93d518"} Oct 14 15:31:04 crc kubenswrapper[4945]: I1014 15:31:04.886490 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"afed9c20-1964-44dd-a849-af9fc0f97e86","Type":"ContainerStarted","Data":"99bca69f011d5f39ffcb04b66ef2e136b1bc5ebcf0fcd85a79cad3f958ed79e2"} Oct 14 15:31:04 crc kubenswrapper[4945]: I1014 15:31:04.889625 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f0fa2603-a8a2-449d-88d2-e35975a18ad1","Type":"ContainerStarted","Data":"a1652df746fc8257c6e83adc82327efa89b5ad75c48e5e4662afb68e9395eb2f"} Oct 14 15:31:04 crc kubenswrapper[4945]: I1014 15:31:04.917143 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=23.964672003 podStartE2EDuration="1m3.917117945s" podCreationTimestamp="2025-10-14 15:30:01 +0000 UTC" firstStartedPulling="2025-10-14 15:30:17.63974357 +0000 UTC m=+947.623791938" lastFinishedPulling="2025-10-14 15:30:57.592189492 +0000 UTC m=+987.576237880" observedRunningTime="2025-10-14 15:31:04.910779994 +0000 UTC m=+994.894828372" watchObservedRunningTime="2025-10-14 15:31:04.917117945 +0000 UTC m=+994.901166313" Oct 14 15:31:04 crc kubenswrapper[4945]: I1014 15:31:04.942381 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=21.432559975 podStartE2EDuration="1m1.942360915s" podCreationTimestamp="2025-10-14 15:30:03 +0000 UTC" firstStartedPulling="2025-10-14 15:30:17.106807599 +0000 UTC m=+947.090855977" lastFinishedPulling="2025-10-14 15:30:57.616608549 +0000 UTC m=+987.600656917" observedRunningTime="2025-10-14 15:31:04.938644269 +0000 UTC m=+994.922692637" watchObservedRunningTime="2025-10-14 15:31:04.942360915 +0000 UTC m=+994.926409283" Oct 14 15:31:05 crc kubenswrapper[4945]: I1014 15:31:05.825231 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:31:05 crc kubenswrapper[4945]: E1014 15:31:05.825358 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 15:31:05 crc kubenswrapper[4945]: E1014 15:31:05.825373 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 15:31:05 crc kubenswrapper[4945]: E1014 15:31:05.825412 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift podName:eb8ab92b-e774-40b3-becd-458958582bef nodeName:}" failed. No retries permitted until 2025-10-14 15:31:13.825399792 +0000 UTC m=+1003.809448150 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift") pod "swift-storage-0" (UID: "eb8ab92b-e774-40b3-becd-458958582bef") : configmap "swift-ring-files" not found Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.211869 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.262491 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mfqx8"] Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.263028 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" podUID="39764f61-839c-4425-9b8b-13e92865b7d2" containerName="dnsmasq-dns" containerID="cri-o://aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff" gracePeriod=10 Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.650813 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.763132 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39764f61-839c-4425-9b8b-13e92865b7d2-dns-svc\") pod \"39764f61-839c-4425-9b8b-13e92865b7d2\" (UID: \"39764f61-839c-4425-9b8b-13e92865b7d2\") " Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.763213 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39764f61-839c-4425-9b8b-13e92865b7d2-config\") pod \"39764f61-839c-4425-9b8b-13e92865b7d2\" (UID: \"39764f61-839c-4425-9b8b-13e92865b7d2\") " Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.763355 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhfgw\" (UniqueName: \"kubernetes.io/projected/39764f61-839c-4425-9b8b-13e92865b7d2-kube-api-access-mhfgw\") pod \"39764f61-839c-4425-9b8b-13e92865b7d2\" (UID: \"39764f61-839c-4425-9b8b-13e92865b7d2\") " Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.769080 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39764f61-839c-4425-9b8b-13e92865b7d2-kube-api-access-mhfgw" (OuterVolumeSpecName: "kube-api-access-mhfgw") pod "39764f61-839c-4425-9b8b-13e92865b7d2" (UID: "39764f61-839c-4425-9b8b-13e92865b7d2"). InnerVolumeSpecName "kube-api-access-mhfgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.797792 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39764f61-839c-4425-9b8b-13e92865b7d2-config" (OuterVolumeSpecName: "config") pod "39764f61-839c-4425-9b8b-13e92865b7d2" (UID: "39764f61-839c-4425-9b8b-13e92865b7d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.816785 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39764f61-839c-4425-9b8b-13e92865b7d2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "39764f61-839c-4425-9b8b-13e92865b7d2" (UID: "39764f61-839c-4425-9b8b-13e92865b7d2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.865902 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39764f61-839c-4425-9b8b-13e92865b7d2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.866221 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39764f61-839c-4425-9b8b-13e92865b7d2-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.866303 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhfgw\" (UniqueName: \"kubernetes.io/projected/39764f61-839c-4425-9b8b-13e92865b7d2-kube-api-access-mhfgw\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.913979 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-h5b5w" event={"ID":"787ea529-5875-4e6c-8a3e-5b4fbf019fa8","Type":"ContainerStarted","Data":"d7e9d1b585478d87f8fdf5492d93477196acc9ea2803ec4a24cde153015f6895"} Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.916471 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31","Type":"ContainerStarted","Data":"7c2f5f775c840ff1967dba967a57d6320075403005785c2cdc4bd85935b5eac7"} Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.916500 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31","Type":"ContainerStarted","Data":"458a1cd031b946531875f3882d013b2f72f3f5ce65af0a3cd73c9908aa7a1dde"} Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.917077 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.929768 4945 generic.go:334] "Generic (PLEG): container finished" podID="39764f61-839c-4425-9b8b-13e92865b7d2" containerID="aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff" exitCode=0 Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.929868 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" event={"ID":"39764f61-839c-4425-9b8b-13e92865b7d2","Type":"ContainerDied","Data":"aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff"} Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.930129 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" event={"ID":"39764f61-839c-4425-9b8b-13e92865b7d2","Type":"ContainerDied","Data":"ea6fdfc8b1a1f31e5aab49d6a135b032bdc2a9fea7835d1283911f56990a79cc"} Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.930214 4945 scope.go:117] "RemoveContainer" containerID="aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.929933 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mfqx8" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.933500 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-h5b5w" podStartSLOduration=2.8555327950000002 podStartE2EDuration="6.933484501s" podCreationTimestamp="2025-10-14 15:31:01 +0000 UTC" firstStartedPulling="2025-10-14 15:31:03.058051068 +0000 UTC m=+993.042099436" lastFinishedPulling="2025-10-14 15:31:07.136002774 +0000 UTC m=+997.120051142" observedRunningTime="2025-10-14 15:31:07.932479343 +0000 UTC m=+997.916527751" watchObservedRunningTime="2025-10-14 15:31:07.933484501 +0000 UTC m=+997.917532869" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.968107 4945 scope.go:117] "RemoveContainer" containerID="1fd88d1780df49018ae4a6358b2aa595893e0f5ee39d30f6da12278e064dda68" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.973150 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.729541304 podStartE2EDuration="4.973129852s" podCreationTimestamp="2025-10-14 15:31:03 +0000 UTC" firstStartedPulling="2025-10-14 15:31:03.835989757 +0000 UTC m=+993.820038125" lastFinishedPulling="2025-10-14 15:31:07.079578305 +0000 UTC m=+997.063626673" observedRunningTime="2025-10-14 15:31:07.95553529 +0000 UTC m=+997.939583648" watchObservedRunningTime="2025-10-14 15:31:07.973129852 +0000 UTC m=+997.957178230" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.983062 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mfqx8"] Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.988043 4945 scope.go:117] "RemoveContainer" containerID="aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff" Oct 14 15:31:07 crc kubenswrapper[4945]: E1014 15:31:07.990187 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff\": container with ID starting with aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff not found: ID does not exist" containerID="aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.990232 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff"} err="failed to get container status \"aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff\": rpc error: code = NotFound desc = could not find container \"aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff\": container with ID starting with aaa63604e0befe63bb03c03c43a875cb2b082c785e8d3651264055dff5dcffff not found: ID does not exist" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.990261 4945 scope.go:117] "RemoveContainer" containerID="1fd88d1780df49018ae4a6358b2aa595893e0f5ee39d30f6da12278e064dda68" Oct 14 15:31:07 crc kubenswrapper[4945]: E1014 15:31:07.990565 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fd88d1780df49018ae4a6358b2aa595893e0f5ee39d30f6da12278e064dda68\": container with ID starting with 1fd88d1780df49018ae4a6358b2aa595893e0f5ee39d30f6da12278e064dda68 not found: ID does not exist" containerID="1fd88d1780df49018ae4a6358b2aa595893e0f5ee39d30f6da12278e064dda68" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.990638 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fd88d1780df49018ae4a6358b2aa595893e0f5ee39d30f6da12278e064dda68"} err="failed to get container status \"1fd88d1780df49018ae4a6358b2aa595893e0f5ee39d30f6da12278e064dda68\": rpc error: code = NotFound desc = could not find container \"1fd88d1780df49018ae4a6358b2aa595893e0f5ee39d30f6da12278e064dda68\": container with ID starting with 1fd88d1780df49018ae4a6358b2aa595893e0f5ee39d30f6da12278e064dda68 not found: ID does not exist" Oct 14 15:31:07 crc kubenswrapper[4945]: I1014 15:31:07.992133 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mfqx8"] Oct 14 15:31:08 crc kubenswrapper[4945]: I1014 15:31:08.771441 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39764f61-839c-4425-9b8b-13e92865b7d2" path="/var/lib/kubelet/pods/39764f61-839c-4425-9b8b-13e92865b7d2/volumes" Oct 14 15:31:11 crc kubenswrapper[4945]: I1014 15:31:11.265381 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6v9cw" podUID="9d0edbee-18bd-41ac-880b-526477ef54d3" containerName="ovn-controller" probeResult="failure" output=< Oct 14 15:31:11 crc kubenswrapper[4945]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 14 15:31:11 crc kubenswrapper[4945]: > Oct 14 15:31:12 crc kubenswrapper[4945]: I1014 15:31:12.947681 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 14 15:31:12 crc kubenswrapper[4945]: I1014 15:31:12.947766 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 14 15:31:12 crc kubenswrapper[4945]: I1014 15:31:12.994028 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 14 15:31:13 crc kubenswrapper[4945]: I1014 15:31:13.038305 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 14 15:31:13 crc kubenswrapper[4945]: I1014 15:31:13.865301 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:31:13 crc kubenswrapper[4945]: E1014 15:31:13.865508 4945 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 15:31:13 crc kubenswrapper[4945]: E1014 15:31:13.865535 4945 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 15:31:13 crc kubenswrapper[4945]: E1014 15:31:13.865593 4945 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift podName:eb8ab92b-e774-40b3-becd-458958582bef nodeName:}" failed. No retries permitted until 2025-10-14 15:31:29.865575986 +0000 UTC m=+1019.849624354 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift") pod "swift-storage-0" (UID: "eb8ab92b-e774-40b3-becd-458958582bef") : configmap "swift-ring-files" not found Oct 14 15:31:13 crc kubenswrapper[4945]: I1014 15:31:13.977190 4945 generic.go:334] "Generic (PLEG): container finished" podID="9642a3f8-eb9b-4f79-8448-602a5f7761d6" containerID="421c62871dd88c2e8493610a2bf3454c6bdaab0cc2d9530739b1bc7a8d4d5e01" exitCode=0 Oct 14 15:31:13 crc kubenswrapper[4945]: I1014 15:31:13.977260 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9642a3f8-eb9b-4f79-8448-602a5f7761d6","Type":"ContainerDied","Data":"421c62871dd88c2e8493610a2bf3454c6bdaab0cc2d9530739b1bc7a8d4d5e01"} Oct 14 15:31:13 crc kubenswrapper[4945]: I1014 15:31:13.979080 4945 generic.go:334] "Generic (PLEG): container finished" podID="c7915803-bfbc-4150-854d-6039c685a771" containerID="9c60658f6099b1c229bfd7f9e639bfeb03e6915d14caaeb7e9f7520dc1e46c01" exitCode=0 Oct 14 15:31:13 crc kubenswrapper[4945]: I1014 15:31:13.979176 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c7915803-bfbc-4150-854d-6039c685a771","Type":"ContainerDied","Data":"9c60658f6099b1c229bfd7f9e639bfeb03e6915d14caaeb7e9f7520dc1e46c01"} Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.270069 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-m6rmj"] Oct 14 15:31:14 crc kubenswrapper[4945]: E1014 15:31:14.271011 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39764f61-839c-4425-9b8b-13e92865b7d2" containerName="dnsmasq-dns" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.271599 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="39764f61-839c-4425-9b8b-13e92865b7d2" containerName="dnsmasq-dns" Oct 14 15:31:14 crc kubenswrapper[4945]: E1014 15:31:14.271698 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39764f61-839c-4425-9b8b-13e92865b7d2" containerName="init" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.271758 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="39764f61-839c-4425-9b8b-13e92865b7d2" containerName="init" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.271995 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="39764f61-839c-4425-9b8b-13e92865b7d2" containerName="dnsmasq-dns" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.272550 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-m6rmj" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.281856 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-m6rmj"] Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.374167 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbsm5\" (UniqueName: \"kubernetes.io/projected/c47e35eb-9ce0-4ce8-9b41-defa17743ccc-kube-api-access-bbsm5\") pod \"keystone-db-create-m6rmj\" (UID: \"c47e35eb-9ce0-4ce8-9b41-defa17743ccc\") " pod="openstack/keystone-db-create-m6rmj" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.384648 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.384693 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.432230 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.475655 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbsm5\" (UniqueName: \"kubernetes.io/projected/c47e35eb-9ce0-4ce8-9b41-defa17743ccc-kube-api-access-bbsm5\") pod \"keystone-db-create-m6rmj\" (UID: \"c47e35eb-9ce0-4ce8-9b41-defa17743ccc\") " pod="openstack/keystone-db-create-m6rmj" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.505855 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbsm5\" (UniqueName: \"kubernetes.io/projected/c47e35eb-9ce0-4ce8-9b41-defa17743ccc-kube-api-access-bbsm5\") pod \"keystone-db-create-m6rmj\" (UID: \"c47e35eb-9ce0-4ce8-9b41-defa17743ccc\") " pod="openstack/keystone-db-create-m6rmj" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.508314 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-hmwc5"] Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.509246 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hmwc5" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.524009 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-hmwc5"] Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.577511 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr794\" (UniqueName: \"kubernetes.io/projected/a886bcc8-a794-45d1-a4f4-90ffabf9fb58-kube-api-access-zr794\") pod \"placement-db-create-hmwc5\" (UID: \"a886bcc8-a794-45d1-a4f4-90ffabf9fb58\") " pod="openstack/placement-db-create-hmwc5" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.587377 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-m6rmj" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.679014 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr794\" (UniqueName: \"kubernetes.io/projected/a886bcc8-a794-45d1-a4f4-90ffabf9fb58-kube-api-access-zr794\") pod \"placement-db-create-hmwc5\" (UID: \"a886bcc8-a794-45d1-a4f4-90ffabf9fb58\") " pod="openstack/placement-db-create-hmwc5" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.701732 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr794\" (UniqueName: \"kubernetes.io/projected/a886bcc8-a794-45d1-a4f4-90ffabf9fb58-kube-api-access-zr794\") pod \"placement-db-create-hmwc5\" (UID: \"a886bcc8-a794-45d1-a4f4-90ffabf9fb58\") " pod="openstack/placement-db-create-hmwc5" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.872321 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hmwc5" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.990184 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c7915803-bfbc-4150-854d-6039c685a771","Type":"ContainerStarted","Data":"18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753"} Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.991424 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.993822 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9642a3f8-eb9b-4f79-8448-602a5f7761d6","Type":"ContainerStarted","Data":"ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96"} Oct 14 15:31:14 crc kubenswrapper[4945]: I1014 15:31:14.994504 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:31:15 crc kubenswrapper[4945]: I1014 15:31:15.004780 4945 generic.go:334] "Generic (PLEG): container finished" podID="787ea529-5875-4e6c-8a3e-5b4fbf019fa8" containerID="d7e9d1b585478d87f8fdf5492d93477196acc9ea2803ec4a24cde153015f6895" exitCode=0 Oct 14 15:31:15 crc kubenswrapper[4945]: I1014 15:31:15.006218 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-h5b5w" event={"ID":"787ea529-5875-4e6c-8a3e-5b4fbf019fa8","Type":"ContainerDied","Data":"d7e9d1b585478d87f8fdf5492d93477196acc9ea2803ec4a24cde153015f6895"} Oct 14 15:31:15 crc kubenswrapper[4945]: I1014 15:31:15.021623 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=56.451704347 podStartE2EDuration="1m16.021602729s" podCreationTimestamp="2025-10-14 15:29:59 +0000 UTC" firstStartedPulling="2025-10-14 15:30:18.313305283 +0000 UTC m=+948.297353651" lastFinishedPulling="2025-10-14 15:30:37.883203665 +0000 UTC m=+967.867252033" observedRunningTime="2025-10-14 15:31:15.015110384 +0000 UTC m=+1004.999158772" watchObservedRunningTime="2025-10-14 15:31:15.021602729 +0000 UTC m=+1005.005651097" Oct 14 15:31:15 crc kubenswrapper[4945]: I1014 15:31:15.078280 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-m6rmj"] Oct 14 15:31:15 crc kubenswrapper[4945]: I1014 15:31:15.086486 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=55.443739317 podStartE2EDuration="1m16.086463409s" podCreationTimestamp="2025-10-14 15:29:59 +0000 UTC" firstStartedPulling="2025-10-14 15:30:17.643164888 +0000 UTC m=+947.627213256" lastFinishedPulling="2025-10-14 15:30:38.28588898 +0000 UTC m=+968.269937348" observedRunningTime="2025-10-14 15:31:15.067961001 +0000 UTC m=+1005.052009369" watchObservedRunningTime="2025-10-14 15:31:15.086463409 +0000 UTC m=+1005.070511777" Oct 14 15:31:15 crc kubenswrapper[4945]: I1014 15:31:15.103705 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 14 15:31:15 crc kubenswrapper[4945]: I1014 15:31:15.328545 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-hmwc5"] Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.015031 4945 generic.go:334] "Generic (PLEG): container finished" podID="c47e35eb-9ce0-4ce8-9b41-defa17743ccc" containerID="5df90a957d6f8c9f7a3786d257950ae5fd5cf6972ecd9a1dc2a77b1fa4847c1a" exitCode=0 Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.015126 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-m6rmj" event={"ID":"c47e35eb-9ce0-4ce8-9b41-defa17743ccc","Type":"ContainerDied","Data":"5df90a957d6f8c9f7a3786d257950ae5fd5cf6972ecd9a1dc2a77b1fa4847c1a"} Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.015388 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-m6rmj" event={"ID":"c47e35eb-9ce0-4ce8-9b41-defa17743ccc","Type":"ContainerStarted","Data":"ebf62de234ea20971b648a9235b3c2f7fe1740ac8cf5fce255f093c2aaf1e3b1"} Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.016671 4945 generic.go:334] "Generic (PLEG): container finished" podID="a886bcc8-a794-45d1-a4f4-90ffabf9fb58" containerID="29f039448f9d5fab5df0aa15d63401bfe7a6c6bb55809f04f1e34935e25835ce" exitCode=0 Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.016715 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hmwc5" event={"ID":"a886bcc8-a794-45d1-a4f4-90ffabf9fb58","Type":"ContainerDied","Data":"29f039448f9d5fab5df0aa15d63401bfe7a6c6bb55809f04f1e34935e25835ce"} Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.016761 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hmwc5" event={"ID":"a886bcc8-a794-45d1-a4f4-90ffabf9fb58","Type":"ContainerStarted","Data":"96778ddaac004902366b9771da449ecff26082a3293769d0667d2df1608cd95f"} Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.275062 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6v9cw" podUID="9d0edbee-18bd-41ac-880b-526477ef54d3" containerName="ovn-controller" probeResult="failure" output=< Oct 14 15:31:16 crc kubenswrapper[4945]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 14 15:31:16 crc kubenswrapper[4945]: > Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.293237 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.308516 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-pq8xf" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.400809 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.511124 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-ring-data-devices\") pod \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.511182 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-swiftconf\") pod \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.511210 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-etc-swift\") pod \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.511249 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-dispersionconf\") pod \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.511330 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsjwh\" (UniqueName: \"kubernetes.io/projected/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-kube-api-access-fsjwh\") pod \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.511414 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-combined-ca-bundle\") pod \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.511442 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-scripts\") pod \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\" (UID: \"787ea529-5875-4e6c-8a3e-5b4fbf019fa8\") " Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.512418 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "787ea529-5875-4e6c-8a3e-5b4fbf019fa8" (UID: "787ea529-5875-4e6c-8a3e-5b4fbf019fa8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.512561 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "787ea529-5875-4e6c-8a3e-5b4fbf019fa8" (UID: "787ea529-5875-4e6c-8a3e-5b4fbf019fa8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.533465 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "787ea529-5875-4e6c-8a3e-5b4fbf019fa8" (UID: "787ea529-5875-4e6c-8a3e-5b4fbf019fa8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.542589 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6v9cw-config-pqc4j"] Oct 14 15:31:16 crc kubenswrapper[4945]: E1014 15:31:16.543027 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="787ea529-5875-4e6c-8a3e-5b4fbf019fa8" containerName="swift-ring-rebalance" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.543056 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="787ea529-5875-4e6c-8a3e-5b4fbf019fa8" containerName="swift-ring-rebalance" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.546397 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-kube-api-access-fsjwh" (OuterVolumeSpecName: "kube-api-access-fsjwh") pod "787ea529-5875-4e6c-8a3e-5b4fbf019fa8" (UID: "787ea529-5875-4e6c-8a3e-5b4fbf019fa8"). InnerVolumeSpecName "kube-api-access-fsjwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.547270 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="787ea529-5875-4e6c-8a3e-5b4fbf019fa8" containerName="swift-ring-rebalance" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.548025 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.550775 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "787ea529-5875-4e6c-8a3e-5b4fbf019fa8" (UID: "787ea529-5875-4e6c-8a3e-5b4fbf019fa8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.551066 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.551080 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "787ea529-5875-4e6c-8a3e-5b4fbf019fa8" (UID: "787ea529-5875-4e6c-8a3e-5b4fbf019fa8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.553157 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6v9cw-config-pqc4j"] Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.574370 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-scripts" (OuterVolumeSpecName: "scripts") pod "787ea529-5875-4e6c-8a3e-5b4fbf019fa8" (UID: "787ea529-5875-4e6c-8a3e-5b4fbf019fa8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.613262 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-log-ovn\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.613583 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gscfm\" (UniqueName: \"kubernetes.io/projected/e07c4483-3d4c-4ade-959d-7733f3767348-kube-api-access-gscfm\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.613660 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-run\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.613686 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e07c4483-3d4c-4ade-959d-7733f3767348-additional-scripts\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.613770 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-run-ovn\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.613825 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e07c4483-3d4c-4ade-959d-7733f3767348-scripts\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.613924 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsjwh\" (UniqueName: \"kubernetes.io/projected/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-kube-api-access-fsjwh\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.613940 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.613953 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.613966 4945 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.613978 4945 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.613990 4945 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.614000 4945 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/787ea529-5875-4e6c-8a3e-5b4fbf019fa8-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.715106 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e07c4483-3d4c-4ade-959d-7733f3767348-scripts\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.715237 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-log-ovn\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.715555 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-log-ovn\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.715265 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gscfm\" (UniqueName: \"kubernetes.io/projected/e07c4483-3d4c-4ade-959d-7733f3767348-kube-api-access-gscfm\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.715703 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-run\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.715733 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e07c4483-3d4c-4ade-959d-7733f3767348-additional-scripts\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.715800 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-run\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.716455 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e07c4483-3d4c-4ade-959d-7733f3767348-additional-scripts\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.716616 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-run-ovn\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.716704 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-run-ovn\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.717452 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e07c4483-3d4c-4ade-959d-7733f3767348-scripts\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.738866 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gscfm\" (UniqueName: \"kubernetes.io/projected/e07c4483-3d4c-4ade-959d-7733f3767348-kube-api-access-gscfm\") pod \"ovn-controller-6v9cw-config-pqc4j\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.796336 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.796406 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:31:16 crc kubenswrapper[4945]: I1014 15:31:16.961046 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:17 crc kubenswrapper[4945]: I1014 15:31:17.026092 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-h5b5w" Oct 14 15:31:17 crc kubenswrapper[4945]: I1014 15:31:17.026777 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-h5b5w" event={"ID":"787ea529-5875-4e6c-8a3e-5b4fbf019fa8","Type":"ContainerDied","Data":"b3fe26bf6d5bdba7358eb3a61a7037e1712ab5b4c8e45524920de8fb1a93d518"} Oct 14 15:31:17 crc kubenswrapper[4945]: I1014 15:31:17.026800 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3fe26bf6d5bdba7358eb3a61a7037e1712ab5b4c8e45524920de8fb1a93d518" Oct 14 15:31:17 crc kubenswrapper[4945]: I1014 15:31:17.462233 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hmwc5" Oct 14 15:31:17 crc kubenswrapper[4945]: I1014 15:31:17.468112 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-m6rmj" Oct 14 15:31:17 crc kubenswrapper[4945]: I1014 15:31:17.535552 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr794\" (UniqueName: \"kubernetes.io/projected/a886bcc8-a794-45d1-a4f4-90ffabf9fb58-kube-api-access-zr794\") pod \"a886bcc8-a794-45d1-a4f4-90ffabf9fb58\" (UID: \"a886bcc8-a794-45d1-a4f4-90ffabf9fb58\") " Oct 14 15:31:17 crc kubenswrapper[4945]: I1014 15:31:17.535684 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbsm5\" (UniqueName: \"kubernetes.io/projected/c47e35eb-9ce0-4ce8-9b41-defa17743ccc-kube-api-access-bbsm5\") pod \"c47e35eb-9ce0-4ce8-9b41-defa17743ccc\" (UID: \"c47e35eb-9ce0-4ce8-9b41-defa17743ccc\") " Oct 14 15:31:17 crc kubenswrapper[4945]: I1014 15:31:17.543824 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a886bcc8-a794-45d1-a4f4-90ffabf9fb58-kube-api-access-zr794" (OuterVolumeSpecName: "kube-api-access-zr794") pod "a886bcc8-a794-45d1-a4f4-90ffabf9fb58" (UID: "a886bcc8-a794-45d1-a4f4-90ffabf9fb58"). InnerVolumeSpecName "kube-api-access-zr794". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:17 crc kubenswrapper[4945]: I1014 15:31:17.547187 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c47e35eb-9ce0-4ce8-9b41-defa17743ccc-kube-api-access-bbsm5" (OuterVolumeSpecName: "kube-api-access-bbsm5") pod "c47e35eb-9ce0-4ce8-9b41-defa17743ccc" (UID: "c47e35eb-9ce0-4ce8-9b41-defa17743ccc"). InnerVolumeSpecName "kube-api-access-bbsm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:17 crc kubenswrapper[4945]: I1014 15:31:17.567837 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6v9cw-config-pqc4j"] Oct 14 15:31:17 crc kubenswrapper[4945]: I1014 15:31:17.638675 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr794\" (UniqueName: \"kubernetes.io/projected/a886bcc8-a794-45d1-a4f4-90ffabf9fb58-kube-api-access-zr794\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:17 crc kubenswrapper[4945]: I1014 15:31:17.639036 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbsm5\" (UniqueName: \"kubernetes.io/projected/c47e35eb-9ce0-4ce8-9b41-defa17743ccc-kube-api-access-bbsm5\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:18 crc kubenswrapper[4945]: I1014 15:31:18.035151 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hmwc5" event={"ID":"a886bcc8-a794-45d1-a4f4-90ffabf9fb58","Type":"ContainerDied","Data":"96778ddaac004902366b9771da449ecff26082a3293769d0667d2df1608cd95f"} Oct 14 15:31:18 crc kubenswrapper[4945]: I1014 15:31:18.035188 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hmwc5" Oct 14 15:31:18 crc kubenswrapper[4945]: I1014 15:31:18.035195 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96778ddaac004902366b9771da449ecff26082a3293769d0667d2df1608cd95f" Oct 14 15:31:18 crc kubenswrapper[4945]: I1014 15:31:18.036462 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6v9cw-config-pqc4j" event={"ID":"e07c4483-3d4c-4ade-959d-7733f3767348","Type":"ContainerStarted","Data":"029e0e8b46bea97eb2a68ddc28e4fe0d550539f0ef84ffbd4a6a66b034850e04"} Oct 14 15:31:18 crc kubenswrapper[4945]: I1014 15:31:18.036490 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6v9cw-config-pqc4j" event={"ID":"e07c4483-3d4c-4ade-959d-7733f3767348","Type":"ContainerStarted","Data":"6a77dd9414b0d915972700aa0a3546c5f6909db85578f837164df25444342d88"} Oct 14 15:31:18 crc kubenswrapper[4945]: I1014 15:31:18.037619 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-m6rmj" event={"ID":"c47e35eb-9ce0-4ce8-9b41-defa17743ccc","Type":"ContainerDied","Data":"ebf62de234ea20971b648a9235b3c2f7fe1740ac8cf5fce255f093c2aaf1e3b1"} Oct 14 15:31:18 crc kubenswrapper[4945]: I1014 15:31:18.037638 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebf62de234ea20971b648a9235b3c2f7fe1740ac8cf5fce255f093c2aaf1e3b1" Oct 14 15:31:18 crc kubenswrapper[4945]: I1014 15:31:18.037942 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-m6rmj" Oct 14 15:31:18 crc kubenswrapper[4945]: I1014 15:31:18.058044 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6v9cw-config-pqc4j" podStartSLOduration=2.058029098 podStartE2EDuration="2.058029098s" podCreationTimestamp="2025-10-14 15:31:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:31:18.054760695 +0000 UTC m=+1008.038809143" watchObservedRunningTime="2025-10-14 15:31:18.058029098 +0000 UTC m=+1008.042077466" Oct 14 15:31:18 crc kubenswrapper[4945]: I1014 15:31:18.458218 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 14 15:31:19 crc kubenswrapper[4945]: I1014 15:31:19.045489 4945 generic.go:334] "Generic (PLEG): container finished" podID="e07c4483-3d4c-4ade-959d-7733f3767348" containerID="029e0e8b46bea97eb2a68ddc28e4fe0d550539f0ef84ffbd4a6a66b034850e04" exitCode=0 Oct 14 15:31:19 crc kubenswrapper[4945]: I1014 15:31:19.045533 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6v9cw-config-pqc4j" event={"ID":"e07c4483-3d4c-4ade-959d-7733f3767348","Type":"ContainerDied","Data":"029e0e8b46bea97eb2a68ddc28e4fe0d550539f0ef84ffbd4a6a66b034850e04"} Oct 14 15:31:19 crc kubenswrapper[4945]: I1014 15:31:19.777481 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-kknvq"] Oct 14 15:31:19 crc kubenswrapper[4945]: E1014 15:31:19.777812 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c47e35eb-9ce0-4ce8-9b41-defa17743ccc" containerName="mariadb-database-create" Oct 14 15:31:19 crc kubenswrapper[4945]: I1014 15:31:19.777832 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c47e35eb-9ce0-4ce8-9b41-defa17743ccc" containerName="mariadb-database-create" Oct 14 15:31:19 crc kubenswrapper[4945]: E1014 15:31:19.777851 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a886bcc8-a794-45d1-a4f4-90ffabf9fb58" containerName="mariadb-database-create" Oct 14 15:31:19 crc kubenswrapper[4945]: I1014 15:31:19.777858 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a886bcc8-a794-45d1-a4f4-90ffabf9fb58" containerName="mariadb-database-create" Oct 14 15:31:19 crc kubenswrapper[4945]: I1014 15:31:19.778104 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c47e35eb-9ce0-4ce8-9b41-defa17743ccc" containerName="mariadb-database-create" Oct 14 15:31:19 crc kubenswrapper[4945]: I1014 15:31:19.778130 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a886bcc8-a794-45d1-a4f4-90ffabf9fb58" containerName="mariadb-database-create" Oct 14 15:31:19 crc kubenswrapper[4945]: I1014 15:31:19.778659 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kknvq" Oct 14 15:31:19 crc kubenswrapper[4945]: I1014 15:31:19.790104 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kknvq"] Oct 14 15:31:19 crc kubenswrapper[4945]: I1014 15:31:19.873641 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b4qc\" (UniqueName: \"kubernetes.io/projected/cf94dd02-5151-404b-8444-24a1cb34ca64-kube-api-access-5b4qc\") pod \"glance-db-create-kknvq\" (UID: \"cf94dd02-5151-404b-8444-24a1cb34ca64\") " pod="openstack/glance-db-create-kknvq" Oct 14 15:31:19 crc kubenswrapper[4945]: I1014 15:31:19.975108 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b4qc\" (UniqueName: \"kubernetes.io/projected/cf94dd02-5151-404b-8444-24a1cb34ca64-kube-api-access-5b4qc\") pod \"glance-db-create-kknvq\" (UID: \"cf94dd02-5151-404b-8444-24a1cb34ca64\") " pod="openstack/glance-db-create-kknvq" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.012682 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b4qc\" (UniqueName: \"kubernetes.io/projected/cf94dd02-5151-404b-8444-24a1cb34ca64-kube-api-access-5b4qc\") pod \"glance-db-create-kknvq\" (UID: \"cf94dd02-5151-404b-8444-24a1cb34ca64\") " pod="openstack/glance-db-create-kknvq" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.110771 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kknvq" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.386099 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.482301 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e07c4483-3d4c-4ade-959d-7733f3767348-additional-scripts\") pod \"e07c4483-3d4c-4ade-959d-7733f3767348\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.482373 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e07c4483-3d4c-4ade-959d-7733f3767348-scripts\") pod \"e07c4483-3d4c-4ade-959d-7733f3767348\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.482455 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-run-ovn\") pod \"e07c4483-3d4c-4ade-959d-7733f3767348\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.482486 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gscfm\" (UniqueName: \"kubernetes.io/projected/e07c4483-3d4c-4ade-959d-7733f3767348-kube-api-access-gscfm\") pod \"e07c4483-3d4c-4ade-959d-7733f3767348\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.482511 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-log-ovn\") pod \"e07c4483-3d4c-4ade-959d-7733f3767348\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.482578 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-run\") pod \"e07c4483-3d4c-4ade-959d-7733f3767348\" (UID: \"e07c4483-3d4c-4ade-959d-7733f3767348\") " Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.482578 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e07c4483-3d4c-4ade-959d-7733f3767348" (UID: "e07c4483-3d4c-4ade-959d-7733f3767348"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.482601 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e07c4483-3d4c-4ade-959d-7733f3767348" (UID: "e07c4483-3d4c-4ade-959d-7733f3767348"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.482622 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-run" (OuterVolumeSpecName: "var-run") pod "e07c4483-3d4c-4ade-959d-7733f3767348" (UID: "e07c4483-3d4c-4ade-959d-7733f3767348"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.482937 4945 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.482950 4945 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.482958 4945 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e07c4483-3d4c-4ade-959d-7733f3767348-var-run\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.483375 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e07c4483-3d4c-4ade-959d-7733f3767348-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e07c4483-3d4c-4ade-959d-7733f3767348" (UID: "e07c4483-3d4c-4ade-959d-7733f3767348"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.483629 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e07c4483-3d4c-4ade-959d-7733f3767348-scripts" (OuterVolumeSpecName: "scripts") pod "e07c4483-3d4c-4ade-959d-7733f3767348" (UID: "e07c4483-3d4c-4ade-959d-7733f3767348"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.487661 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e07c4483-3d4c-4ade-959d-7733f3767348-kube-api-access-gscfm" (OuterVolumeSpecName: "kube-api-access-gscfm") pod "e07c4483-3d4c-4ade-959d-7733f3767348" (UID: "e07c4483-3d4c-4ade-959d-7733f3767348"). InnerVolumeSpecName "kube-api-access-gscfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.582407 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kknvq"] Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.584928 4945 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e07c4483-3d4c-4ade-959d-7733f3767348-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.584956 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e07c4483-3d4c-4ade-959d-7733f3767348-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.584965 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gscfm\" (UniqueName: \"kubernetes.io/projected/e07c4483-3d4c-4ade-959d-7733f3767348-kube-api-access-gscfm\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:20 crc kubenswrapper[4945]: W1014 15:31:20.586115 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf94dd02_5151_404b_8444_24a1cb34ca64.slice/crio-0a9815a6abfc6b1d3b405e25672b72aa155f3303a4377f4e51eeaeb1f2912ced WatchSource:0}: Error finding container 0a9815a6abfc6b1d3b405e25672b72aa155f3303a4377f4e51eeaeb1f2912ced: Status 404 returned error can't find the container with id 0a9815a6abfc6b1d3b405e25672b72aa155f3303a4377f4e51eeaeb1f2912ced Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.662581 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-6v9cw-config-pqc4j"] Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.668910 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-6v9cw-config-pqc4j"] Oct 14 15:31:20 crc kubenswrapper[4945]: I1014 15:31:20.773865 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e07c4483-3d4c-4ade-959d-7733f3767348" path="/var/lib/kubelet/pods/e07c4483-3d4c-4ade-959d-7733f3767348/volumes" Oct 14 15:31:21 crc kubenswrapper[4945]: I1014 15:31:21.060685 4945 scope.go:117] "RemoveContainer" containerID="029e0e8b46bea97eb2a68ddc28e4fe0d550539f0ef84ffbd4a6a66b034850e04" Oct 14 15:31:21 crc kubenswrapper[4945]: I1014 15:31:21.060685 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6v9cw-config-pqc4j" Oct 14 15:31:21 crc kubenswrapper[4945]: I1014 15:31:21.062629 4945 generic.go:334] "Generic (PLEG): container finished" podID="cf94dd02-5151-404b-8444-24a1cb34ca64" containerID="cb47933954470f605e8b5db33fc6d09991dd6f92a7fd05fa837dee287180cba3" exitCode=0 Oct 14 15:31:21 crc kubenswrapper[4945]: I1014 15:31:21.062657 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kknvq" event={"ID":"cf94dd02-5151-404b-8444-24a1cb34ca64","Type":"ContainerDied","Data":"cb47933954470f605e8b5db33fc6d09991dd6f92a7fd05fa837dee287180cba3"} Oct 14 15:31:21 crc kubenswrapper[4945]: I1014 15:31:21.062675 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kknvq" event={"ID":"cf94dd02-5151-404b-8444-24a1cb34ca64","Type":"ContainerStarted","Data":"0a9815a6abfc6b1d3b405e25672b72aa155f3303a4377f4e51eeaeb1f2912ced"} Oct 14 15:31:21 crc kubenswrapper[4945]: I1014 15:31:21.267460 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-6v9cw" Oct 14 15:31:22 crc kubenswrapper[4945]: I1014 15:31:22.416811 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kknvq" Oct 14 15:31:22 crc kubenswrapper[4945]: I1014 15:31:22.513370 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5b4qc\" (UniqueName: \"kubernetes.io/projected/cf94dd02-5151-404b-8444-24a1cb34ca64-kube-api-access-5b4qc\") pod \"cf94dd02-5151-404b-8444-24a1cb34ca64\" (UID: \"cf94dd02-5151-404b-8444-24a1cb34ca64\") " Oct 14 15:31:22 crc kubenswrapper[4945]: I1014 15:31:22.523674 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf94dd02-5151-404b-8444-24a1cb34ca64-kube-api-access-5b4qc" (OuterVolumeSpecName: "kube-api-access-5b4qc") pod "cf94dd02-5151-404b-8444-24a1cb34ca64" (UID: "cf94dd02-5151-404b-8444-24a1cb34ca64"). InnerVolumeSpecName "kube-api-access-5b4qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:22 crc kubenswrapper[4945]: I1014 15:31:22.615339 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5b4qc\" (UniqueName: \"kubernetes.io/projected/cf94dd02-5151-404b-8444-24a1cb34ca64-kube-api-access-5b4qc\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:23 crc kubenswrapper[4945]: I1014 15:31:23.085837 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kknvq" event={"ID":"cf94dd02-5151-404b-8444-24a1cb34ca64","Type":"ContainerDied","Data":"0a9815a6abfc6b1d3b405e25672b72aa155f3303a4377f4e51eeaeb1f2912ced"} Oct 14 15:31:23 crc kubenswrapper[4945]: I1014 15:31:23.085907 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a9815a6abfc6b1d3b405e25672b72aa155f3303a4377f4e51eeaeb1f2912ced" Oct 14 15:31:23 crc kubenswrapper[4945]: I1014 15:31:23.085962 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kknvq" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.292439 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-812a-account-create-cf984"] Oct 14 15:31:24 crc kubenswrapper[4945]: E1014 15:31:24.293232 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e07c4483-3d4c-4ade-959d-7733f3767348" containerName="ovn-config" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.293244 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e07c4483-3d4c-4ade-959d-7733f3767348" containerName="ovn-config" Oct 14 15:31:24 crc kubenswrapper[4945]: E1014 15:31:24.293262 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf94dd02-5151-404b-8444-24a1cb34ca64" containerName="mariadb-database-create" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.293268 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf94dd02-5151-404b-8444-24a1cb34ca64" containerName="mariadb-database-create" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.293428 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf94dd02-5151-404b-8444-24a1cb34ca64" containerName="mariadb-database-create" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.293448 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e07c4483-3d4c-4ade-959d-7733f3767348" containerName="ovn-config" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.294028 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-812a-account-create-cf984" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.296726 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.302771 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-812a-account-create-cf984"] Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.448801 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq2xs\" (UniqueName: \"kubernetes.io/projected/7e394f9f-f797-4659-b933-432c563567d8-kube-api-access-mq2xs\") pod \"keystone-812a-account-create-cf984\" (UID: \"7e394f9f-f797-4659-b933-432c563567d8\") " pod="openstack/keystone-812a-account-create-cf984" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.550400 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq2xs\" (UniqueName: \"kubernetes.io/projected/7e394f9f-f797-4659-b933-432c563567d8-kube-api-access-mq2xs\") pod \"keystone-812a-account-create-cf984\" (UID: \"7e394f9f-f797-4659-b933-432c563567d8\") " pod="openstack/keystone-812a-account-create-cf984" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.573106 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq2xs\" (UniqueName: \"kubernetes.io/projected/7e394f9f-f797-4659-b933-432c563567d8-kube-api-access-mq2xs\") pod \"keystone-812a-account-create-cf984\" (UID: \"7e394f9f-f797-4659-b933-432c563567d8\") " pod="openstack/keystone-812a-account-create-cf984" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.650412 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-812a-account-create-cf984" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.718296 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-ab43-account-create-sksdh"] Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.719987 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ab43-account-create-sksdh" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.722236 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.727522 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ab43-account-create-sksdh"] Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.859993 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqmgh\" (UniqueName: \"kubernetes.io/projected/51f09adf-1049-4876-b658-4552d03dc683-kube-api-access-wqmgh\") pod \"placement-ab43-account-create-sksdh\" (UID: \"51f09adf-1049-4876-b658-4552d03dc683\") " pod="openstack/placement-ab43-account-create-sksdh" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.962030 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqmgh\" (UniqueName: \"kubernetes.io/projected/51f09adf-1049-4876-b658-4552d03dc683-kube-api-access-wqmgh\") pod \"placement-ab43-account-create-sksdh\" (UID: \"51f09adf-1049-4876-b658-4552d03dc683\") " pod="openstack/placement-ab43-account-create-sksdh" Oct 14 15:31:24 crc kubenswrapper[4945]: I1014 15:31:24.980397 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqmgh\" (UniqueName: \"kubernetes.io/projected/51f09adf-1049-4876-b658-4552d03dc683-kube-api-access-wqmgh\") pod \"placement-ab43-account-create-sksdh\" (UID: \"51f09adf-1049-4876-b658-4552d03dc683\") " pod="openstack/placement-ab43-account-create-sksdh" Oct 14 15:31:25 crc kubenswrapper[4945]: I1014 15:31:25.069206 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ab43-account-create-sksdh" Oct 14 15:31:25 crc kubenswrapper[4945]: I1014 15:31:25.108713 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-812a-account-create-cf984"] Oct 14 15:31:25 crc kubenswrapper[4945]: W1014 15:31:25.118241 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e394f9f_f797_4659_b933_432c563567d8.slice/crio-830c49b82c8dd4bfb071a92a69bb60397f8dc33bf96a8d09567b85e1d745bbbb WatchSource:0}: Error finding container 830c49b82c8dd4bfb071a92a69bb60397f8dc33bf96a8d09567b85e1d745bbbb: Status 404 returned error can't find the container with id 830c49b82c8dd4bfb071a92a69bb60397f8dc33bf96a8d09567b85e1d745bbbb Oct 14 15:31:25 crc kubenswrapper[4945]: I1014 15:31:25.499014 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ab43-account-create-sksdh"] Oct 14 15:31:25 crc kubenswrapper[4945]: W1014 15:31:25.502674 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51f09adf_1049_4876_b658_4552d03dc683.slice/crio-5f3938293de1edd46d9736c911f57ff0af75298bb12926a59f936f881bad2e36 WatchSource:0}: Error finding container 5f3938293de1edd46d9736c911f57ff0af75298bb12926a59f936f881bad2e36: Status 404 returned error can't find the container with id 5f3938293de1edd46d9736c911f57ff0af75298bb12926a59f936f881bad2e36 Oct 14 15:31:26 crc kubenswrapper[4945]: I1014 15:31:26.115590 4945 generic.go:334] "Generic (PLEG): container finished" podID="51f09adf-1049-4876-b658-4552d03dc683" containerID="906273dd7ab06c572849d59c59594ae665e8df032223b846c25cd5fe75e2ab8b" exitCode=0 Oct 14 15:31:26 crc kubenswrapper[4945]: I1014 15:31:26.115699 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ab43-account-create-sksdh" event={"ID":"51f09adf-1049-4876-b658-4552d03dc683","Type":"ContainerDied","Data":"906273dd7ab06c572849d59c59594ae665e8df032223b846c25cd5fe75e2ab8b"} Oct 14 15:31:26 crc kubenswrapper[4945]: I1014 15:31:26.115742 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ab43-account-create-sksdh" event={"ID":"51f09adf-1049-4876-b658-4552d03dc683","Type":"ContainerStarted","Data":"5f3938293de1edd46d9736c911f57ff0af75298bb12926a59f936f881bad2e36"} Oct 14 15:31:26 crc kubenswrapper[4945]: I1014 15:31:26.117020 4945 generic.go:334] "Generic (PLEG): container finished" podID="7e394f9f-f797-4659-b933-432c563567d8" containerID="32c05eacfa9988f89fc32fd69945888db53d27247a1244a098b0ce8e2aaaaf43" exitCode=0 Oct 14 15:31:26 crc kubenswrapper[4945]: I1014 15:31:26.117053 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-812a-account-create-cf984" event={"ID":"7e394f9f-f797-4659-b933-432c563567d8","Type":"ContainerDied","Data":"32c05eacfa9988f89fc32fd69945888db53d27247a1244a098b0ce8e2aaaaf43"} Oct 14 15:31:26 crc kubenswrapper[4945]: I1014 15:31:26.117089 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-812a-account-create-cf984" event={"ID":"7e394f9f-f797-4659-b933-432c563567d8","Type":"ContainerStarted","Data":"830c49b82c8dd4bfb071a92a69bb60397f8dc33bf96a8d09567b85e1d745bbbb"} Oct 14 15:31:27 crc kubenswrapper[4945]: I1014 15:31:27.471677 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ab43-account-create-sksdh" Oct 14 15:31:27 crc kubenswrapper[4945]: I1014 15:31:27.477418 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-812a-account-create-cf984" Oct 14 15:31:27 crc kubenswrapper[4945]: I1014 15:31:27.605112 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq2xs\" (UniqueName: \"kubernetes.io/projected/7e394f9f-f797-4659-b933-432c563567d8-kube-api-access-mq2xs\") pod \"7e394f9f-f797-4659-b933-432c563567d8\" (UID: \"7e394f9f-f797-4659-b933-432c563567d8\") " Oct 14 15:31:27 crc kubenswrapper[4945]: I1014 15:31:27.605310 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqmgh\" (UniqueName: \"kubernetes.io/projected/51f09adf-1049-4876-b658-4552d03dc683-kube-api-access-wqmgh\") pod \"51f09adf-1049-4876-b658-4552d03dc683\" (UID: \"51f09adf-1049-4876-b658-4552d03dc683\") " Oct 14 15:31:27 crc kubenswrapper[4945]: I1014 15:31:27.610440 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e394f9f-f797-4659-b933-432c563567d8-kube-api-access-mq2xs" (OuterVolumeSpecName: "kube-api-access-mq2xs") pod "7e394f9f-f797-4659-b933-432c563567d8" (UID: "7e394f9f-f797-4659-b933-432c563567d8"). InnerVolumeSpecName "kube-api-access-mq2xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:27 crc kubenswrapper[4945]: I1014 15:31:27.613110 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51f09adf-1049-4876-b658-4552d03dc683-kube-api-access-wqmgh" (OuterVolumeSpecName: "kube-api-access-wqmgh") pod "51f09adf-1049-4876-b658-4552d03dc683" (UID: "51f09adf-1049-4876-b658-4552d03dc683"). InnerVolumeSpecName "kube-api-access-wqmgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:27 crc kubenswrapper[4945]: I1014 15:31:27.707594 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq2xs\" (UniqueName: \"kubernetes.io/projected/7e394f9f-f797-4659-b933-432c563567d8-kube-api-access-mq2xs\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:27 crc kubenswrapper[4945]: I1014 15:31:27.707841 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqmgh\" (UniqueName: \"kubernetes.io/projected/51f09adf-1049-4876-b658-4552d03dc683-kube-api-access-wqmgh\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:28 crc kubenswrapper[4945]: I1014 15:31:28.132105 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-812a-account-create-cf984" event={"ID":"7e394f9f-f797-4659-b933-432c563567d8","Type":"ContainerDied","Data":"830c49b82c8dd4bfb071a92a69bb60397f8dc33bf96a8d09567b85e1d745bbbb"} Oct 14 15:31:28 crc kubenswrapper[4945]: I1014 15:31:28.132152 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="830c49b82c8dd4bfb071a92a69bb60397f8dc33bf96a8d09567b85e1d745bbbb" Oct 14 15:31:28 crc kubenswrapper[4945]: I1014 15:31:28.132117 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-812a-account-create-cf984" Oct 14 15:31:28 crc kubenswrapper[4945]: I1014 15:31:28.134170 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ab43-account-create-sksdh" event={"ID":"51f09adf-1049-4876-b658-4552d03dc683","Type":"ContainerDied","Data":"5f3938293de1edd46d9736c911f57ff0af75298bb12926a59f936f881bad2e36"} Oct 14 15:31:28 crc kubenswrapper[4945]: I1014 15:31:28.134196 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f3938293de1edd46d9736c911f57ff0af75298bb12926a59f936f881bad2e36" Oct 14 15:31:28 crc kubenswrapper[4945]: I1014 15:31:28.134214 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ab43-account-create-sksdh" Oct 14 15:31:29 crc kubenswrapper[4945]: I1014 15:31:29.836906 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-534f-account-create-nl9hz"] Oct 14 15:31:29 crc kubenswrapper[4945]: E1014 15:31:29.837270 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51f09adf-1049-4876-b658-4552d03dc683" containerName="mariadb-account-create" Oct 14 15:31:29 crc kubenswrapper[4945]: I1014 15:31:29.837284 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="51f09adf-1049-4876-b658-4552d03dc683" containerName="mariadb-account-create" Oct 14 15:31:29 crc kubenswrapper[4945]: E1014 15:31:29.837294 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e394f9f-f797-4659-b933-432c563567d8" containerName="mariadb-account-create" Oct 14 15:31:29 crc kubenswrapper[4945]: I1014 15:31:29.837300 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e394f9f-f797-4659-b933-432c563567d8" containerName="mariadb-account-create" Oct 14 15:31:29 crc kubenswrapper[4945]: I1014 15:31:29.837515 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="51f09adf-1049-4876-b658-4552d03dc683" containerName="mariadb-account-create" Oct 14 15:31:29 crc kubenswrapper[4945]: I1014 15:31:29.837528 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e394f9f-f797-4659-b933-432c563567d8" containerName="mariadb-account-create" Oct 14 15:31:29 crc kubenswrapper[4945]: I1014 15:31:29.838100 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-534f-account-create-nl9hz" Oct 14 15:31:29 crc kubenswrapper[4945]: I1014 15:31:29.842887 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 14 15:31:29 crc kubenswrapper[4945]: I1014 15:31:29.845797 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-534f-account-create-nl9hz"] Oct 14 15:31:29 crc kubenswrapper[4945]: I1014 15:31:29.944201 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzxl9\" (UniqueName: \"kubernetes.io/projected/eed35ce9-830a-4746-a5f7-e2e7cfd44d75-kube-api-access-hzxl9\") pod \"glance-534f-account-create-nl9hz\" (UID: \"eed35ce9-830a-4746-a5f7-e2e7cfd44d75\") " pod="openstack/glance-534f-account-create-nl9hz" Oct 14 15:31:29 crc kubenswrapper[4945]: I1014 15:31:29.944480 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:31:29 crc kubenswrapper[4945]: I1014 15:31:29.961769 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/eb8ab92b-e774-40b3-becd-458958582bef-etc-swift\") pod \"swift-storage-0\" (UID: \"eb8ab92b-e774-40b3-becd-458958582bef\") " pod="openstack/swift-storage-0" Oct 14 15:31:30 crc kubenswrapper[4945]: I1014 15:31:30.045738 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzxl9\" (UniqueName: \"kubernetes.io/projected/eed35ce9-830a-4746-a5f7-e2e7cfd44d75-kube-api-access-hzxl9\") pod \"glance-534f-account-create-nl9hz\" (UID: \"eed35ce9-830a-4746-a5f7-e2e7cfd44d75\") " pod="openstack/glance-534f-account-create-nl9hz" Oct 14 15:31:30 crc kubenswrapper[4945]: I1014 15:31:30.068551 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzxl9\" (UniqueName: \"kubernetes.io/projected/eed35ce9-830a-4746-a5f7-e2e7cfd44d75-kube-api-access-hzxl9\") pod \"glance-534f-account-create-nl9hz\" (UID: \"eed35ce9-830a-4746-a5f7-e2e7cfd44d75\") " pod="openstack/glance-534f-account-create-nl9hz" Oct 14 15:31:30 crc kubenswrapper[4945]: I1014 15:31:30.115801 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 14 15:31:30 crc kubenswrapper[4945]: I1014 15:31:30.162677 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-534f-account-create-nl9hz" Oct 14 15:31:30 crc kubenswrapper[4945]: I1014 15:31:30.671597 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-534f-account-create-nl9hz"] Oct 14 15:31:30 crc kubenswrapper[4945]: W1014 15:31:30.678509 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeed35ce9_830a_4746_a5f7_e2e7cfd44d75.slice/crio-9ec1f1a29ba665699cf77eae2ae83f603957639e52206081b7a857fcd89e3323 WatchSource:0}: Error finding container 9ec1f1a29ba665699cf77eae2ae83f603957639e52206081b7a857fcd89e3323: Status 404 returned error can't find the container with id 9ec1f1a29ba665699cf77eae2ae83f603957639e52206081b7a857fcd89e3323 Oct 14 15:31:30 crc kubenswrapper[4945]: I1014 15:31:30.678951 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 14 15:31:30 crc kubenswrapper[4945]: W1014 15:31:30.687691 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb8ab92b_e774_40b3_becd_458958582bef.slice/crio-5db8c4f20bf3c94c014878f6997ef83efb9a0e4302bc2d88cfd00871ff2fb1f6 WatchSource:0}: Error finding container 5db8c4f20bf3c94c014878f6997ef83efb9a0e4302bc2d88cfd00871ff2fb1f6: Status 404 returned error can't find the container with id 5db8c4f20bf3c94c014878f6997ef83efb9a0e4302bc2d88cfd00871ff2fb1f6 Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.158460 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"5db8c4f20bf3c94c014878f6997ef83efb9a0e4302bc2d88cfd00871ff2fb1f6"} Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.161247 4945 generic.go:334] "Generic (PLEG): container finished" podID="eed35ce9-830a-4746-a5f7-e2e7cfd44d75" containerID="887307eb130aeee4a121c651b4244748e960cb1684871a6e58db2aa322bcec1c" exitCode=0 Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.161384 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-534f-account-create-nl9hz" event={"ID":"eed35ce9-830a-4746-a5f7-e2e7cfd44d75","Type":"ContainerDied","Data":"887307eb130aeee4a121c651b4244748e960cb1684871a6e58db2aa322bcec1c"} Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.161493 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-534f-account-create-nl9hz" event={"ID":"eed35ce9-830a-4746-a5f7-e2e7cfd44d75","Type":"ContainerStarted","Data":"9ec1f1a29ba665699cf77eae2ae83f603957639e52206081b7a857fcd89e3323"} Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.339098 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.466132 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.708063 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-7cz6h"] Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.709301 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-7cz6h" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.727380 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-7cz6h"] Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.771343 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwpc7\" (UniqueName: \"kubernetes.io/projected/826c5051-a7b9-4f9a-98dd-cb1954e00a10-kube-api-access-vwpc7\") pod \"cinder-db-create-7cz6h\" (UID: \"826c5051-a7b9-4f9a-98dd-cb1954e00a10\") " pod="openstack/cinder-db-create-7cz6h" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.810231 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-9k955"] Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.811700 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-9k955" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.820664 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-9k955"] Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.872694 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ndjx\" (UniqueName: \"kubernetes.io/projected/1c9d06f9-312e-464b-9438-4f2f34a4daf0-kube-api-access-4ndjx\") pod \"barbican-db-create-9k955\" (UID: \"1c9d06f9-312e-464b-9438-4f2f34a4daf0\") " pod="openstack/barbican-db-create-9k955" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.872871 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwpc7\" (UniqueName: \"kubernetes.io/projected/826c5051-a7b9-4f9a-98dd-cb1954e00a10-kube-api-access-vwpc7\") pod \"cinder-db-create-7cz6h\" (UID: \"826c5051-a7b9-4f9a-98dd-cb1954e00a10\") " pod="openstack/cinder-db-create-7cz6h" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.893388 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwpc7\" (UniqueName: \"kubernetes.io/projected/826c5051-a7b9-4f9a-98dd-cb1954e00a10-kube-api-access-vwpc7\") pod \"cinder-db-create-7cz6h\" (UID: \"826c5051-a7b9-4f9a-98dd-cb1954e00a10\") " pod="openstack/cinder-db-create-7cz6h" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.910105 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-tfzfz"] Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.921205 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tfzfz"] Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.921316 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tfzfz" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.974152 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ndjx\" (UniqueName: \"kubernetes.io/projected/1c9d06f9-312e-464b-9438-4f2f34a4daf0-kube-api-access-4ndjx\") pod \"barbican-db-create-9k955\" (UID: \"1c9d06f9-312e-464b-9438-4f2f34a4daf0\") " pod="openstack/barbican-db-create-9k955" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.981779 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-d6w8h"] Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.982806 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.987256 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.987583 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-g69v8" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.989152 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.989391 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 15:31:31 crc kubenswrapper[4945]: I1014 15:31:31.999624 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-d6w8h"] Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.011483 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ndjx\" (UniqueName: \"kubernetes.io/projected/1c9d06f9-312e-464b-9438-4f2f34a4daf0-kube-api-access-4ndjx\") pod \"barbican-db-create-9k955\" (UID: \"1c9d06f9-312e-464b-9438-4f2f34a4daf0\") " pod="openstack/barbican-db-create-9k955" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.062154 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-7cz6h" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.076028 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2bpt\" (UniqueName: \"kubernetes.io/projected/887c4eba-a780-4417-a255-528cfd654d8b-kube-api-access-c2bpt\") pod \"neutron-db-create-tfzfz\" (UID: \"887c4eba-a780-4417-a255-528cfd654d8b\") " pod="openstack/neutron-db-create-tfzfz" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.076072 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c11ab49-aef7-4d60-b287-147be1668631-combined-ca-bundle\") pod \"keystone-db-sync-d6w8h\" (UID: \"1c11ab49-aef7-4d60-b287-147be1668631\") " pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.076120 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c11ab49-aef7-4d60-b287-147be1668631-config-data\") pod \"keystone-db-sync-d6w8h\" (UID: \"1c11ab49-aef7-4d60-b287-147be1668631\") " pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.076321 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjszg\" (UniqueName: \"kubernetes.io/projected/1c11ab49-aef7-4d60-b287-147be1668631-kube-api-access-pjszg\") pod \"keystone-db-sync-d6w8h\" (UID: \"1c11ab49-aef7-4d60-b287-147be1668631\") " pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.142904 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-9k955" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.177621 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjszg\" (UniqueName: \"kubernetes.io/projected/1c11ab49-aef7-4d60-b287-147be1668631-kube-api-access-pjszg\") pod \"keystone-db-sync-d6w8h\" (UID: \"1c11ab49-aef7-4d60-b287-147be1668631\") " pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.177681 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2bpt\" (UniqueName: \"kubernetes.io/projected/887c4eba-a780-4417-a255-528cfd654d8b-kube-api-access-c2bpt\") pod \"neutron-db-create-tfzfz\" (UID: \"887c4eba-a780-4417-a255-528cfd654d8b\") " pod="openstack/neutron-db-create-tfzfz" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.177709 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c11ab49-aef7-4d60-b287-147be1668631-combined-ca-bundle\") pod \"keystone-db-sync-d6w8h\" (UID: \"1c11ab49-aef7-4d60-b287-147be1668631\") " pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.177759 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c11ab49-aef7-4d60-b287-147be1668631-config-data\") pod \"keystone-db-sync-d6w8h\" (UID: \"1c11ab49-aef7-4d60-b287-147be1668631\") " pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.181417 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c11ab49-aef7-4d60-b287-147be1668631-combined-ca-bundle\") pod \"keystone-db-sync-d6w8h\" (UID: \"1c11ab49-aef7-4d60-b287-147be1668631\") " pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.182140 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c11ab49-aef7-4d60-b287-147be1668631-config-data\") pod \"keystone-db-sync-d6w8h\" (UID: \"1c11ab49-aef7-4d60-b287-147be1668631\") " pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.196688 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2bpt\" (UniqueName: \"kubernetes.io/projected/887c4eba-a780-4417-a255-528cfd654d8b-kube-api-access-c2bpt\") pod \"neutron-db-create-tfzfz\" (UID: \"887c4eba-a780-4417-a255-528cfd654d8b\") " pod="openstack/neutron-db-create-tfzfz" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.200490 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjszg\" (UniqueName: \"kubernetes.io/projected/1c11ab49-aef7-4d60-b287-147be1668631-kube-api-access-pjszg\") pod \"keystone-db-sync-d6w8h\" (UID: \"1c11ab49-aef7-4d60-b287-147be1668631\") " pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.256194 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tfzfz" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.306920 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.511423 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-534f-account-create-nl9hz" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.584259 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzxl9\" (UniqueName: \"kubernetes.io/projected/eed35ce9-830a-4746-a5f7-e2e7cfd44d75-kube-api-access-hzxl9\") pod \"eed35ce9-830a-4746-a5f7-e2e7cfd44d75\" (UID: \"eed35ce9-830a-4746-a5f7-e2e7cfd44d75\") " Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.588079 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eed35ce9-830a-4746-a5f7-e2e7cfd44d75-kube-api-access-hzxl9" (OuterVolumeSpecName: "kube-api-access-hzxl9") pod "eed35ce9-830a-4746-a5f7-e2e7cfd44d75" (UID: "eed35ce9-830a-4746-a5f7-e2e7cfd44d75"). InnerVolumeSpecName "kube-api-access-hzxl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.687565 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzxl9\" (UniqueName: \"kubernetes.io/projected/eed35ce9-830a-4746-a5f7-e2e7cfd44d75-kube-api-access-hzxl9\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:32 crc kubenswrapper[4945]: I1014 15:31:32.840842 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-9k955"] Oct 14 15:31:33 crc kubenswrapper[4945]: I1014 15:31:33.123656 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-7cz6h"] Oct 14 15:31:33 crc kubenswrapper[4945]: I1014 15:31:33.136767 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-d6w8h"] Oct 14 15:31:33 crc kubenswrapper[4945]: I1014 15:31:33.146548 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tfzfz"] Oct 14 15:31:33 crc kubenswrapper[4945]: I1014 15:31:33.185231 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-7cz6h" event={"ID":"826c5051-a7b9-4f9a-98dd-cb1954e00a10","Type":"ContainerStarted","Data":"f52ade35b1ccb981c3da49c664a028eaa5727cc02c645d38f6f4dc452a907e19"} Oct 14 15:31:33 crc kubenswrapper[4945]: I1014 15:31:33.191957 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"96ec54407cd2e33d0f7b35b8a8445775af0592ba682bb58165a0395336157572"} Oct 14 15:31:33 crc kubenswrapper[4945]: I1014 15:31:33.191997 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"c7f0257e089cc63e0ade77056744ff420741f996b82812ec671809798fd544fd"} Oct 14 15:31:33 crc kubenswrapper[4945]: I1014 15:31:33.192009 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"2f45fb18b720e3d13b6e33113cc0f8711185e2208f0a1ef5ea0a214997cc0b67"} Oct 14 15:31:33 crc kubenswrapper[4945]: I1014 15:31:33.195949 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-9k955" event={"ID":"1c9d06f9-312e-464b-9438-4f2f34a4daf0","Type":"ContainerStarted","Data":"61af3dd7ce199caf3ff5f9ca2275d347c15460ce7d65f14b36456335f3de13ed"} Oct 14 15:31:33 crc kubenswrapper[4945]: I1014 15:31:33.195992 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-9k955" event={"ID":"1c9d06f9-312e-464b-9438-4f2f34a4daf0","Type":"ContainerStarted","Data":"4dec05e365ca0237bf95eadf4dca49b12f4e0421205195eded0ad96a10c5b6b8"} Oct 14 15:31:33 crc kubenswrapper[4945]: I1014 15:31:33.216640 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-534f-account-create-nl9hz" event={"ID":"eed35ce9-830a-4746-a5f7-e2e7cfd44d75","Type":"ContainerDied","Data":"9ec1f1a29ba665699cf77eae2ae83f603957639e52206081b7a857fcd89e3323"} Oct 14 15:31:33 crc kubenswrapper[4945]: I1014 15:31:33.216687 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ec1f1a29ba665699cf77eae2ae83f603957639e52206081b7a857fcd89e3323" Oct 14 15:31:33 crc kubenswrapper[4945]: I1014 15:31:33.217477 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-534f-account-create-nl9hz" Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.232135 4945 generic.go:334] "Generic (PLEG): container finished" podID="887c4eba-a780-4417-a255-528cfd654d8b" containerID="8b09de0a388a4af4b207025b85d07804726d41e9b3a99c8f0e1578134df48dc1" exitCode=0 Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.232469 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tfzfz" event={"ID":"887c4eba-a780-4417-a255-528cfd654d8b","Type":"ContainerDied","Data":"8b09de0a388a4af4b207025b85d07804726d41e9b3a99c8f0e1578134df48dc1"} Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.232503 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tfzfz" event={"ID":"887c4eba-a780-4417-a255-528cfd654d8b","Type":"ContainerStarted","Data":"067e78b2bc21a31e2e86956dbc89dba8538ae492318635946cee614b2911b0d2"} Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.235338 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-d6w8h" event={"ID":"1c11ab49-aef7-4d60-b287-147be1668631","Type":"ContainerStarted","Data":"b02d7515288c45f0ee09a0b53629f0842a7b838851e451bc6675d9bfba7623f9"} Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.240296 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"187907f6e130e80b1b226800ec3293c9beeebeb3b25dd003be0aa09df5d7d7d7"} Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.242378 4945 generic.go:334] "Generic (PLEG): container finished" podID="1c9d06f9-312e-464b-9438-4f2f34a4daf0" containerID="61af3dd7ce199caf3ff5f9ca2275d347c15460ce7d65f14b36456335f3de13ed" exitCode=0 Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.242433 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-9k955" event={"ID":"1c9d06f9-312e-464b-9438-4f2f34a4daf0","Type":"ContainerDied","Data":"61af3dd7ce199caf3ff5f9ca2275d347c15460ce7d65f14b36456335f3de13ed"} Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.249813 4945 generic.go:334] "Generic (PLEG): container finished" podID="826c5051-a7b9-4f9a-98dd-cb1954e00a10" containerID="4824edc86a1d31578b95250e8420fb346b0eea01d539f2b095e8a2e673d07b5c" exitCode=0 Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.249889 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-7cz6h" event={"ID":"826c5051-a7b9-4f9a-98dd-cb1954e00a10","Type":"ContainerDied","Data":"4824edc86a1d31578b95250e8420fb346b0eea01d539f2b095e8a2e673d07b5c"} Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.673284 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-9k955" Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.834463 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ndjx\" (UniqueName: \"kubernetes.io/projected/1c9d06f9-312e-464b-9438-4f2f34a4daf0-kube-api-access-4ndjx\") pod \"1c9d06f9-312e-464b-9438-4f2f34a4daf0\" (UID: \"1c9d06f9-312e-464b-9438-4f2f34a4daf0\") " Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.841333 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c9d06f9-312e-464b-9438-4f2f34a4daf0-kube-api-access-4ndjx" (OuterVolumeSpecName: "kube-api-access-4ndjx") pod "1c9d06f9-312e-464b-9438-4f2f34a4daf0" (UID: "1c9d06f9-312e-464b-9438-4f2f34a4daf0"). InnerVolumeSpecName "kube-api-access-4ndjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:34 crc kubenswrapper[4945]: I1014 15:31:34.936650 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ndjx\" (UniqueName: \"kubernetes.io/projected/1c9d06f9-312e-464b-9438-4f2f34a4daf0-kube-api-access-4ndjx\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.024399 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-mf7rz"] Oct 14 15:31:35 crc kubenswrapper[4945]: E1014 15:31:35.024788 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eed35ce9-830a-4746-a5f7-e2e7cfd44d75" containerName="mariadb-account-create" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.024811 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="eed35ce9-830a-4746-a5f7-e2e7cfd44d75" containerName="mariadb-account-create" Oct 14 15:31:35 crc kubenswrapper[4945]: E1014 15:31:35.024849 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9d06f9-312e-464b-9438-4f2f34a4daf0" containerName="mariadb-database-create" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.024856 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9d06f9-312e-464b-9438-4f2f34a4daf0" containerName="mariadb-database-create" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.025354 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="eed35ce9-830a-4746-a5f7-e2e7cfd44d75" containerName="mariadb-account-create" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.025388 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c9d06f9-312e-464b-9438-4f2f34a4daf0" containerName="mariadb-database-create" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.025910 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.029001 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-jkrfs" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.029281 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.033394 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mf7rz"] Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.139702 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvg67\" (UniqueName: \"kubernetes.io/projected/bbc6053c-32ba-4e9f-86b7-8590cff47464-kube-api-access-rvg67\") pod \"glance-db-sync-mf7rz\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.139789 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-db-sync-config-data\") pod \"glance-db-sync-mf7rz\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.139867 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-config-data\") pod \"glance-db-sync-mf7rz\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.139910 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-combined-ca-bundle\") pod \"glance-db-sync-mf7rz\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.241545 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-config-data\") pod \"glance-db-sync-mf7rz\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.241607 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-combined-ca-bundle\") pod \"glance-db-sync-mf7rz\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.241697 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvg67\" (UniqueName: \"kubernetes.io/projected/bbc6053c-32ba-4e9f-86b7-8590cff47464-kube-api-access-rvg67\") pod \"glance-db-sync-mf7rz\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.241744 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-db-sync-config-data\") pod \"glance-db-sync-mf7rz\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.246218 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-config-data\") pod \"glance-db-sync-mf7rz\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.248134 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-combined-ca-bundle\") pod \"glance-db-sync-mf7rz\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.248167 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-db-sync-config-data\") pod \"glance-db-sync-mf7rz\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.258699 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvg67\" (UniqueName: \"kubernetes.io/projected/bbc6053c-32ba-4e9f-86b7-8590cff47464-kube-api-access-rvg67\") pod \"glance-db-sync-mf7rz\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.259324 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-9k955" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.260084 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-9k955" event={"ID":"1c9d06f9-312e-464b-9438-4f2f34a4daf0","Type":"ContainerDied","Data":"4dec05e365ca0237bf95eadf4dca49b12f4e0421205195eded0ad96a10c5b6b8"} Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.260137 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4dec05e365ca0237bf95eadf4dca49b12f4e0421205195eded0ad96a10c5b6b8" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.270354 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"909f150c373460c66f557d46ccabc9393050db7a59f00ee8ac8c191c9db84653"} Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.270621 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"e980851526bab0c12453da59da71001126cf74dcec94beeca3a0d12682ab32d8"} Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.270634 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"902161ce8edc9e6981fc1d9c32cf357c73d02f9c0519e363f1aa029336f0290e"} Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.270642 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"faf49931343df5a296ce4dc948511123fd34578d58e98dbd9306cfa7d791cebb"} Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.362610 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mf7rz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.645757 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-7cz6h" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.651202 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tfzfz" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.762382 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2bpt\" (UniqueName: \"kubernetes.io/projected/887c4eba-a780-4417-a255-528cfd654d8b-kube-api-access-c2bpt\") pod \"887c4eba-a780-4417-a255-528cfd654d8b\" (UID: \"887c4eba-a780-4417-a255-528cfd654d8b\") " Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.762572 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwpc7\" (UniqueName: \"kubernetes.io/projected/826c5051-a7b9-4f9a-98dd-cb1954e00a10-kube-api-access-vwpc7\") pod \"826c5051-a7b9-4f9a-98dd-cb1954e00a10\" (UID: \"826c5051-a7b9-4f9a-98dd-cb1954e00a10\") " Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.768693 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/887c4eba-a780-4417-a255-528cfd654d8b-kube-api-access-c2bpt" (OuterVolumeSpecName: "kube-api-access-c2bpt") pod "887c4eba-a780-4417-a255-528cfd654d8b" (UID: "887c4eba-a780-4417-a255-528cfd654d8b"). InnerVolumeSpecName "kube-api-access-c2bpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.773949 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/826c5051-a7b9-4f9a-98dd-cb1954e00a10-kube-api-access-vwpc7" (OuterVolumeSpecName: "kube-api-access-vwpc7") pod "826c5051-a7b9-4f9a-98dd-cb1954e00a10" (UID: "826c5051-a7b9-4f9a-98dd-cb1954e00a10"). InnerVolumeSpecName "kube-api-access-vwpc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.865280 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2bpt\" (UniqueName: \"kubernetes.io/projected/887c4eba-a780-4417-a255-528cfd654d8b-kube-api-access-c2bpt\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.865316 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwpc7\" (UniqueName: \"kubernetes.io/projected/826c5051-a7b9-4f9a-98dd-cb1954e00a10-kube-api-access-vwpc7\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:35 crc kubenswrapper[4945]: I1014 15:31:35.899503 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mf7rz"] Oct 14 15:31:36 crc kubenswrapper[4945]: I1014 15:31:36.318552 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tfzfz" Oct 14 15:31:36 crc kubenswrapper[4945]: I1014 15:31:36.319329 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tfzfz" event={"ID":"887c4eba-a780-4417-a255-528cfd654d8b","Type":"ContainerDied","Data":"067e78b2bc21a31e2e86956dbc89dba8538ae492318635946cee614b2911b0d2"} Oct 14 15:31:36 crc kubenswrapper[4945]: I1014 15:31:36.320733 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="067e78b2bc21a31e2e86956dbc89dba8538ae492318635946cee614b2911b0d2" Oct 14 15:31:36 crc kubenswrapper[4945]: I1014 15:31:36.363309 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-7cz6h" event={"ID":"826c5051-a7b9-4f9a-98dd-cb1954e00a10","Type":"ContainerDied","Data":"f52ade35b1ccb981c3da49c664a028eaa5727cc02c645d38f6f4dc452a907e19"} Oct 14 15:31:36 crc kubenswrapper[4945]: I1014 15:31:36.363354 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f52ade35b1ccb981c3da49c664a028eaa5727cc02c645d38f6f4dc452a907e19" Oct 14 15:31:36 crc kubenswrapper[4945]: I1014 15:31:36.363420 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-7cz6h" Oct 14 15:31:40 crc kubenswrapper[4945]: I1014 15:31:40.409490 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mf7rz" event={"ID":"bbc6053c-32ba-4e9f-86b7-8590cff47464","Type":"ContainerStarted","Data":"a1526d1ea282ac5bcbeb8c8e0690a0e2a50aaf47b3e6592a83551c6a1a77e683"} Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.418072 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-d6w8h" event={"ID":"1c11ab49-aef7-4d60-b287-147be1668631","Type":"ContainerStarted","Data":"7eea8c21fbbf3dd6bcf0bc3b3a1bc387f6d0b95acd41720addf787b8f431ea1d"} Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.758157 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-d6w8h" podStartSLOduration=2.883143277 podStartE2EDuration="10.758133629s" podCreationTimestamp="2025-10-14 15:31:31 +0000 UTC" firstStartedPulling="2025-10-14 15:31:33.18381623 +0000 UTC m=+1023.167864588" lastFinishedPulling="2025-10-14 15:31:41.058806572 +0000 UTC m=+1031.042854940" observedRunningTime="2025-10-14 15:31:41.438951135 +0000 UTC m=+1031.422999513" watchObservedRunningTime="2025-10-14 15:31:41.758133629 +0000 UTC m=+1031.742181997" Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.763023 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-01ac-account-create-c6467"] Oct 14 15:31:41 crc kubenswrapper[4945]: E1014 15:31:41.763465 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="887c4eba-a780-4417-a255-528cfd654d8b" containerName="mariadb-database-create" Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.763482 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="887c4eba-a780-4417-a255-528cfd654d8b" containerName="mariadb-database-create" Oct 14 15:31:41 crc kubenswrapper[4945]: E1014 15:31:41.763504 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826c5051-a7b9-4f9a-98dd-cb1954e00a10" containerName="mariadb-database-create" Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.763512 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="826c5051-a7b9-4f9a-98dd-cb1954e00a10" containerName="mariadb-database-create" Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.763767 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="826c5051-a7b9-4f9a-98dd-cb1954e00a10" containerName="mariadb-database-create" Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.763805 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="887c4eba-a780-4417-a255-528cfd654d8b" containerName="mariadb-database-create" Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.782473 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-01ac-account-create-c6467" Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.791198 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.793101 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-01ac-account-create-c6467"] Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.864146 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62z7j\" (UniqueName: \"kubernetes.io/projected/155850d5-d5d8-419e-8bfb-580ae259b00c-kube-api-access-62z7j\") pod \"barbican-01ac-account-create-c6467\" (UID: \"155850d5-d5d8-419e-8bfb-580ae259b00c\") " pod="openstack/barbican-01ac-account-create-c6467" Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.965524 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62z7j\" (UniqueName: \"kubernetes.io/projected/155850d5-d5d8-419e-8bfb-580ae259b00c-kube-api-access-62z7j\") pod \"barbican-01ac-account-create-c6467\" (UID: \"155850d5-d5d8-419e-8bfb-580ae259b00c\") " pod="openstack/barbican-01ac-account-create-c6467" Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.970468 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-dc8c-account-create-k2465"] Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.971519 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dc8c-account-create-k2465" Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.973302 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.979125 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-dc8c-account-create-k2465"] Oct 14 15:31:41 crc kubenswrapper[4945]: I1014 15:31:41.983693 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62z7j\" (UniqueName: \"kubernetes.io/projected/155850d5-d5d8-419e-8bfb-580ae259b00c-kube-api-access-62z7j\") pod \"barbican-01ac-account-create-c6467\" (UID: \"155850d5-d5d8-419e-8bfb-580ae259b00c\") " pod="openstack/barbican-01ac-account-create-c6467" Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.067133 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9vx4\" (UniqueName: \"kubernetes.io/projected/3ebee178-706b-44af-9df7-d7292d028a28-kube-api-access-c9vx4\") pod \"cinder-dc8c-account-create-k2465\" (UID: \"3ebee178-706b-44af-9df7-d7292d028a28\") " pod="openstack/cinder-dc8c-account-create-k2465" Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.142641 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-01ac-account-create-c6467" Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.168178 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b8b8-account-create-fd6fj"] Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.168584 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9vx4\" (UniqueName: \"kubernetes.io/projected/3ebee178-706b-44af-9df7-d7292d028a28-kube-api-access-c9vx4\") pod \"cinder-dc8c-account-create-k2465\" (UID: \"3ebee178-706b-44af-9df7-d7292d028a28\") " pod="openstack/cinder-dc8c-account-create-k2465" Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.169195 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b8b8-account-create-fd6fj" Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.171280 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.177992 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b8b8-account-create-fd6fj"] Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.190445 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9vx4\" (UniqueName: \"kubernetes.io/projected/3ebee178-706b-44af-9df7-d7292d028a28-kube-api-access-c9vx4\") pod \"cinder-dc8c-account-create-k2465\" (UID: \"3ebee178-706b-44af-9df7-d7292d028a28\") " pod="openstack/cinder-dc8c-account-create-k2465" Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.270216 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc5nv\" (UniqueName: \"kubernetes.io/projected/0c4ea6c9-6e27-4af8-8115-7582222b971f-kube-api-access-lc5nv\") pod \"neutron-b8b8-account-create-fd6fj\" (UID: \"0c4ea6c9-6e27-4af8-8115-7582222b971f\") " pod="openstack/neutron-b8b8-account-create-fd6fj" Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.345698 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dc8c-account-create-k2465" Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.372313 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc5nv\" (UniqueName: \"kubernetes.io/projected/0c4ea6c9-6e27-4af8-8115-7582222b971f-kube-api-access-lc5nv\") pod \"neutron-b8b8-account-create-fd6fj\" (UID: \"0c4ea6c9-6e27-4af8-8115-7582222b971f\") " pod="openstack/neutron-b8b8-account-create-fd6fj" Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.398764 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc5nv\" (UniqueName: \"kubernetes.io/projected/0c4ea6c9-6e27-4af8-8115-7582222b971f-kube-api-access-lc5nv\") pod \"neutron-b8b8-account-create-fd6fj\" (UID: \"0c4ea6c9-6e27-4af8-8115-7582222b971f\") " pod="openstack/neutron-b8b8-account-create-fd6fj" Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.428609 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"0600fc77a94dd941fcbcbb318c8e0639880cfbcaf1b2b7810e353133c61eda6d"} Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.428659 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"3c2d03ad7995cfc711f24b7004e82ae98bd0f317d04369fce9ff759143930024"} Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.428675 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"f4dd34a0973fbe22217bd935265d6907b202cc37389d659367be35e6042ea4ba"} Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.558594 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b8b8-account-create-fd6fj" Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.643524 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-01ac-account-create-c6467"] Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.845499 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-dc8c-account-create-k2465"] Oct 14 15:31:42 crc kubenswrapper[4945]: W1014 15:31:42.862773 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ebee178_706b_44af_9df7_d7292d028a28.slice/crio-779da7c5872ea37642f76683d0f7993f6537440cf8a3ffc0983431ca582589e4 WatchSource:0}: Error finding container 779da7c5872ea37642f76683d0f7993f6537440cf8a3ffc0983431ca582589e4: Status 404 returned error can't find the container with id 779da7c5872ea37642f76683d0f7993f6537440cf8a3ffc0983431ca582589e4 Oct 14 15:31:42 crc kubenswrapper[4945]: I1014 15:31:42.986778 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b8b8-account-create-fd6fj"] Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.444868 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"9e41e329a342d58ab5d83d2ee98097bab7e7d3ab840d27a3c82204f0ef76d553"} Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.444937 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"baa8816512f4aef9abb6f9a51d72f17919ae8d1aae59f1f675be20b696e58646"} Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.444951 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"98bdada999b469c662c7437fa6429cb3f112683c241ecd93afff70c9b121dfe7"} Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.444962 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"eb8ab92b-e774-40b3-becd-458958582bef","Type":"ContainerStarted","Data":"a4b992a2e0e5f267b32ae843fdeac7a333363cbc5b939f43d9bab3770d0e8d1a"} Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.446422 4945 generic.go:334] "Generic (PLEG): container finished" podID="155850d5-d5d8-419e-8bfb-580ae259b00c" containerID="9ea6c19084af198c8763b3c730d8213fd1fbaeb122a274c1c4cf1b805545399d" exitCode=0 Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.446504 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-01ac-account-create-c6467" event={"ID":"155850d5-d5d8-419e-8bfb-580ae259b00c","Type":"ContainerDied","Data":"9ea6c19084af198c8763b3c730d8213fd1fbaeb122a274c1c4cf1b805545399d"} Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.446553 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-01ac-account-create-c6467" event={"ID":"155850d5-d5d8-419e-8bfb-580ae259b00c","Type":"ContainerStarted","Data":"7c98580838fdcb8ed135cdaf13cd72e80e42bb924096cedf6117bf881e334c7b"} Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.448634 4945 generic.go:334] "Generic (PLEG): container finished" podID="3ebee178-706b-44af-9df7-d7292d028a28" containerID="1337f9899efdf21e01d543d6446b2ee1c8e1c275cefd765b667d354f0ad1991c" exitCode=0 Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.448707 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dc8c-account-create-k2465" event={"ID":"3ebee178-706b-44af-9df7-d7292d028a28","Type":"ContainerDied","Data":"1337f9899efdf21e01d543d6446b2ee1c8e1c275cefd765b667d354f0ad1991c"} Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.448741 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dc8c-account-create-k2465" event={"ID":"3ebee178-706b-44af-9df7-d7292d028a28","Type":"ContainerStarted","Data":"779da7c5872ea37642f76683d0f7993f6537440cf8a3ffc0983431ca582589e4"} Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.451511 4945 generic.go:334] "Generic (PLEG): container finished" podID="0c4ea6c9-6e27-4af8-8115-7582222b971f" containerID="760a37312437f03b6331c881c11c4ef944fb136c3556d5ba0f7bd495b2b68dc7" exitCode=0 Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.451547 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b8b8-account-create-fd6fj" event={"ID":"0c4ea6c9-6e27-4af8-8115-7582222b971f","Type":"ContainerDied","Data":"760a37312437f03b6331c881c11c4ef944fb136c3556d5ba0f7bd495b2b68dc7"} Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.451573 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b8b8-account-create-fd6fj" event={"ID":"0c4ea6c9-6e27-4af8-8115-7582222b971f","Type":"ContainerStarted","Data":"d3b0f213969d79bcf044c87b3ca8e69e1615dffd17a4eee545bfb237013db923"} Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.484999 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.496419293 podStartE2EDuration="47.484977604s" podCreationTimestamp="2025-10-14 15:30:56 +0000 UTC" firstStartedPulling="2025-10-14 15:31:30.69497263 +0000 UTC m=+1020.679020998" lastFinishedPulling="2025-10-14 15:31:41.683530941 +0000 UTC m=+1031.667579309" observedRunningTime="2025-10-14 15:31:43.476982406 +0000 UTC m=+1033.461030774" watchObservedRunningTime="2025-10-14 15:31:43.484977604 +0000 UTC m=+1033.469025972" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.739247 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-pdrgh"] Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.740556 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.743385 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.758665 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-pdrgh"] Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.798663 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-config\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.798713 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.798745 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.798760 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.798785 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkb9c\" (UniqueName: \"kubernetes.io/projected/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-kube-api-access-tkb9c\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.798826 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.899740 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.900050 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-config\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.900135 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.900253 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.900356 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.900394 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkb9c\" (UniqueName: \"kubernetes.io/projected/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-kube-api-access-tkb9c\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.900819 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.900951 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.901547 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.901780 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.902705 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-config\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:43 crc kubenswrapper[4945]: I1014 15:31:43.927545 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkb9c\" (UniqueName: \"kubernetes.io/projected/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-kube-api-access-tkb9c\") pod \"dnsmasq-dns-6d5b6d6b67-pdrgh\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:44 crc kubenswrapper[4945]: I1014 15:31:44.088693 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:44 crc kubenswrapper[4945]: I1014 15:31:44.460383 4945 generic.go:334] "Generic (PLEG): container finished" podID="1c11ab49-aef7-4d60-b287-147be1668631" containerID="7eea8c21fbbf3dd6bcf0bc3b3a1bc387f6d0b95acd41720addf787b8f431ea1d" exitCode=0 Oct 14 15:31:44 crc kubenswrapper[4945]: I1014 15:31:44.460479 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-d6w8h" event={"ID":"1c11ab49-aef7-4d60-b287-147be1668631","Type":"ContainerDied","Data":"7eea8c21fbbf3dd6bcf0bc3b3a1bc387f6d0b95acd41720addf787b8f431ea1d"} Oct 14 15:31:44 crc kubenswrapper[4945]: I1014 15:31:44.519327 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-pdrgh"] Oct 14 15:31:44 crc kubenswrapper[4945]: I1014 15:31:44.675389 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dc8c-account-create-k2465" Oct 14 15:31:44 crc kubenswrapper[4945]: I1014 15:31:44.814635 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9vx4\" (UniqueName: \"kubernetes.io/projected/3ebee178-706b-44af-9df7-d7292d028a28-kube-api-access-c9vx4\") pod \"3ebee178-706b-44af-9df7-d7292d028a28\" (UID: \"3ebee178-706b-44af-9df7-d7292d028a28\") " Oct 14 15:31:44 crc kubenswrapper[4945]: I1014 15:31:44.822280 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ebee178-706b-44af-9df7-d7292d028a28-kube-api-access-c9vx4" (OuterVolumeSpecName: "kube-api-access-c9vx4") pod "3ebee178-706b-44af-9df7-d7292d028a28" (UID: "3ebee178-706b-44af-9df7-d7292d028a28"). InnerVolumeSpecName "kube-api-access-c9vx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:44 crc kubenswrapper[4945]: I1014 15:31:44.916916 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9vx4\" (UniqueName: \"kubernetes.io/projected/3ebee178-706b-44af-9df7-d7292d028a28-kube-api-access-c9vx4\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:44 crc kubenswrapper[4945]: I1014 15:31:44.965473 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b8b8-account-create-fd6fj" Oct 14 15:31:44 crc kubenswrapper[4945]: I1014 15:31:44.973227 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-01ac-account-create-c6467" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.017529 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc5nv\" (UniqueName: \"kubernetes.io/projected/0c4ea6c9-6e27-4af8-8115-7582222b971f-kube-api-access-lc5nv\") pod \"0c4ea6c9-6e27-4af8-8115-7582222b971f\" (UID: \"0c4ea6c9-6e27-4af8-8115-7582222b971f\") " Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.023259 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c4ea6c9-6e27-4af8-8115-7582222b971f-kube-api-access-lc5nv" (OuterVolumeSpecName: "kube-api-access-lc5nv") pod "0c4ea6c9-6e27-4af8-8115-7582222b971f" (UID: "0c4ea6c9-6e27-4af8-8115-7582222b971f"). InnerVolumeSpecName "kube-api-access-lc5nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:45 crc kubenswrapper[4945]: E1014 15:31:45.032750 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a1fdf4e_9ca3_4f5d_a5dd_6448afcf7b62.slice/crio-conmon-a69f5155cd00125a9543770689f463e6b67a7eb4a183c43f70c3393133382750.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a1fdf4e_9ca3_4f5d_a5dd_6448afcf7b62.slice/crio-a69f5155cd00125a9543770689f463e6b67a7eb4a183c43f70c3393133382750.scope\": RecentStats: unable to find data in memory cache]" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.120417 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62z7j\" (UniqueName: \"kubernetes.io/projected/155850d5-d5d8-419e-8bfb-580ae259b00c-kube-api-access-62z7j\") pod \"155850d5-d5d8-419e-8bfb-580ae259b00c\" (UID: \"155850d5-d5d8-419e-8bfb-580ae259b00c\") " Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.120932 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc5nv\" (UniqueName: \"kubernetes.io/projected/0c4ea6c9-6e27-4af8-8115-7582222b971f-kube-api-access-lc5nv\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.123160 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/155850d5-d5d8-419e-8bfb-580ae259b00c-kube-api-access-62z7j" (OuterVolumeSpecName: "kube-api-access-62z7j") pod "155850d5-d5d8-419e-8bfb-580ae259b00c" (UID: "155850d5-d5d8-419e-8bfb-580ae259b00c"). InnerVolumeSpecName "kube-api-access-62z7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.222208 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62z7j\" (UniqueName: \"kubernetes.io/projected/155850d5-d5d8-419e-8bfb-580ae259b00c-kube-api-access-62z7j\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.483613 4945 generic.go:334] "Generic (PLEG): container finished" podID="9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" containerID="a69f5155cd00125a9543770689f463e6b67a7eb4a183c43f70c3393133382750" exitCode=0 Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.483718 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" event={"ID":"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62","Type":"ContainerDied","Data":"a69f5155cd00125a9543770689f463e6b67a7eb4a183c43f70c3393133382750"} Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.483749 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" event={"ID":"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62","Type":"ContainerStarted","Data":"2bbad4e701116d32d38127290bb063676c967e6c22ae4d6406a0548f983f9339"} Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.493445 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-01ac-account-create-c6467" event={"ID":"155850d5-d5d8-419e-8bfb-580ae259b00c","Type":"ContainerDied","Data":"7c98580838fdcb8ed135cdaf13cd72e80e42bb924096cedf6117bf881e334c7b"} Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.493485 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c98580838fdcb8ed135cdaf13cd72e80e42bb924096cedf6117bf881e334c7b" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.493558 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-01ac-account-create-c6467" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.495800 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dc8c-account-create-k2465" event={"ID":"3ebee178-706b-44af-9df7-d7292d028a28","Type":"ContainerDied","Data":"779da7c5872ea37642f76683d0f7993f6537440cf8a3ffc0983431ca582589e4"} Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.495842 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="779da7c5872ea37642f76683d0f7993f6537440cf8a3ffc0983431ca582589e4" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.495930 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dc8c-account-create-k2465" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.499307 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b8b8-account-create-fd6fj" event={"ID":"0c4ea6c9-6e27-4af8-8115-7582222b971f","Type":"ContainerDied","Data":"d3b0f213969d79bcf044c87b3ca8e69e1615dffd17a4eee545bfb237013db923"} Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.499351 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b8b8-account-create-fd6fj" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.499376 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3b0f213969d79bcf044c87b3ca8e69e1615dffd17a4eee545bfb237013db923" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.736098 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.832116 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c11ab49-aef7-4d60-b287-147be1668631-combined-ca-bundle\") pod \"1c11ab49-aef7-4d60-b287-147be1668631\" (UID: \"1c11ab49-aef7-4d60-b287-147be1668631\") " Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.832229 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjszg\" (UniqueName: \"kubernetes.io/projected/1c11ab49-aef7-4d60-b287-147be1668631-kube-api-access-pjszg\") pod \"1c11ab49-aef7-4d60-b287-147be1668631\" (UID: \"1c11ab49-aef7-4d60-b287-147be1668631\") " Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.832280 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c11ab49-aef7-4d60-b287-147be1668631-config-data\") pod \"1c11ab49-aef7-4d60-b287-147be1668631\" (UID: \"1c11ab49-aef7-4d60-b287-147be1668631\") " Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.837032 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c11ab49-aef7-4d60-b287-147be1668631-kube-api-access-pjszg" (OuterVolumeSpecName: "kube-api-access-pjszg") pod "1c11ab49-aef7-4d60-b287-147be1668631" (UID: "1c11ab49-aef7-4d60-b287-147be1668631"). InnerVolumeSpecName "kube-api-access-pjszg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.872093 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c11ab49-aef7-4d60-b287-147be1668631-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c11ab49-aef7-4d60-b287-147be1668631" (UID: "1c11ab49-aef7-4d60-b287-147be1668631"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.881300 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c11ab49-aef7-4d60-b287-147be1668631-config-data" (OuterVolumeSpecName: "config-data") pod "1c11ab49-aef7-4d60-b287-147be1668631" (UID: "1c11ab49-aef7-4d60-b287-147be1668631"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.935089 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c11ab49-aef7-4d60-b287-147be1668631-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.935131 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjszg\" (UniqueName: \"kubernetes.io/projected/1c11ab49-aef7-4d60-b287-147be1668631-kube-api-access-pjszg\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:45 crc kubenswrapper[4945]: I1014 15:31:45.935142 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c11ab49-aef7-4d60-b287-147be1668631-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.513279 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" event={"ID":"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62","Type":"ContainerStarted","Data":"86b0bcd60848c10e41aff85b5870d6d5aca2d5b1e9d7649a97b2d78f75c903ab"} Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.513628 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.517177 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-d6w8h" event={"ID":"1c11ab49-aef7-4d60-b287-147be1668631","Type":"ContainerDied","Data":"b02d7515288c45f0ee09a0b53629f0842a7b838851e451bc6675d9bfba7623f9"} Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.517213 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b02d7515288c45f0ee09a0b53629f0842a7b838851e451bc6675d9bfba7623f9" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.517278 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-d6w8h" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.535914 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" podStartSLOduration=3.535897596 podStartE2EDuration="3.535897596s" podCreationTimestamp="2025-10-14 15:31:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:31:46.531951704 +0000 UTC m=+1036.516000082" watchObservedRunningTime="2025-10-14 15:31:46.535897596 +0000 UTC m=+1036.519945964" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.746204 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-pdrgh"] Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.802831 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hgl8j"] Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.803331 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.803383 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:31:46 crc kubenswrapper[4945]: E1014 15:31:46.803349 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ebee178-706b-44af-9df7-d7292d028a28" containerName="mariadb-account-create" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.803501 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ebee178-706b-44af-9df7-d7292d028a28" containerName="mariadb-account-create" Oct 14 15:31:46 crc kubenswrapper[4945]: E1014 15:31:46.803563 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c4ea6c9-6e27-4af8-8115-7582222b971f" containerName="mariadb-account-create" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.803621 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c4ea6c9-6e27-4af8-8115-7582222b971f" containerName="mariadb-account-create" Oct 14 15:31:46 crc kubenswrapper[4945]: E1014 15:31:46.803681 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c11ab49-aef7-4d60-b287-147be1668631" containerName="keystone-db-sync" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.803729 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c11ab49-aef7-4d60-b287-147be1668631" containerName="keystone-db-sync" Oct 14 15:31:46 crc kubenswrapper[4945]: E1014 15:31:46.803797 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155850d5-d5d8-419e-8bfb-580ae259b00c" containerName="mariadb-account-create" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.803853 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="155850d5-d5d8-419e-8bfb-580ae259b00c" containerName="mariadb-account-create" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.804117 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c4ea6c9-6e27-4af8-8115-7582222b971f" containerName="mariadb-account-create" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.804212 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ebee178-706b-44af-9df7-d7292d028a28" containerName="mariadb-account-create" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.804546 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="155850d5-d5d8-419e-8bfb-580ae259b00c" containerName="mariadb-account-create" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.804648 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c11ab49-aef7-4d60-b287-147be1668631" containerName="keystone-db-sync" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.805314 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-bjj42"] Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.806311 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.806668 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-bjj42"] Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.806915 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.815580 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.816899 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.821585 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.822303 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-g69v8" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.829616 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hgl8j"] Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.948912 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-db-create-l5q2x"] Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.950122 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-l5q2x" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.956114 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.956223 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-config-data\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.956253 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.956301 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-credential-keys\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.956326 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-fernet-keys\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.956352 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ngqv\" (UniqueName: \"kubernetes.io/projected/117165e9-bf45-467d-87a8-8237789ca777-kube-api-access-4ngqv\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.956389 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.956424 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-scripts\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.956450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-config\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.956471 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.956491 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-combined-ca-bundle\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.956540 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlhgr\" (UniqueName: \"kubernetes.io/projected/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-kube-api-access-hlhgr\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.968781 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.970824 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.975456 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 15:31:46 crc kubenswrapper[4945]: I1014 15:31:46.982518 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.002229 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.016511 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-create-l5q2x"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058069 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80044b30-c554-4654-a4c0-8ecbcc209d02-run-httpd\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058138 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-scripts\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058175 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-config\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058197 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058223 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-combined-ca-bundle\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058268 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-config-data\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058290 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058332 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlhgr\" (UniqueName: \"kubernetes.io/projected/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-kube-api-access-hlhgr\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058369 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058399 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058451 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6wg7\" (UniqueName: \"kubernetes.io/projected/80044b30-c554-4654-a4c0-8ecbcc209d02-kube-api-access-v6wg7\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058487 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80044b30-c554-4654-a4c0-8ecbcc209d02-log-httpd\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058544 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-config-data\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058569 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058606 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-scripts\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058640 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-credential-keys\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058668 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-fernet-keys\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058693 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gbpf\" (UniqueName: \"kubernetes.io/projected/56ce1365-283f-4f53-8563-60efbce145f7-kube-api-access-2gbpf\") pod \"ironic-db-create-l5q2x\" (UID: \"56ce1365-283f-4f53-8563-60efbce145f7\") " pod="openstack/ironic-db-create-l5q2x" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058718 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ngqv\" (UniqueName: \"kubernetes.io/projected/117165e9-bf45-467d-87a8-8237789ca777-kube-api-access-4ngqv\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.058753 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.059724 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.060819 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-config\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.061130 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.061439 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.062177 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.065506 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-config-data\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.065586 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-fernet-keys\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.066025 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-credential-keys\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.066543 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-scripts\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.068404 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-combined-ca-bundle\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.082766 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlhgr\" (UniqueName: \"kubernetes.io/projected/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-kube-api-access-hlhgr\") pod \"dnsmasq-dns-6f8c45789f-bjj42\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.097242 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ngqv\" (UniqueName: \"kubernetes.io/projected/117165e9-bf45-467d-87a8-8237789ca777-kube-api-access-4ngqv\") pod \"keystone-bootstrap-hgl8j\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.131598 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-bjj42"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.132641 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.134770 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.168933 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-6pgfh"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.171317 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.178422 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80044b30-c554-4654-a4c0-8ecbcc209d02-log-httpd\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.178665 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-scripts\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.178713 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gbpf\" (UniqueName: \"kubernetes.io/projected/56ce1365-283f-4f53-8563-60efbce145f7-kube-api-access-2gbpf\") pod \"ironic-db-create-l5q2x\" (UID: \"56ce1365-283f-4f53-8563-60efbce145f7\") " pod="openstack/ironic-db-create-l5q2x" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.178770 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80044b30-c554-4654-a4c0-8ecbcc209d02-run-httpd\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.189138 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-config-data\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.189421 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.189608 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.189815 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6wg7\" (UniqueName: \"kubernetes.io/projected/80044b30-c554-4654-a4c0-8ecbcc209d02-kube-api-access-v6wg7\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.190960 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-gmlf9" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.197389 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80044b30-c554-4654-a4c0-8ecbcc209d02-log-httpd\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.197471 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80044b30-c554-4654-a4c0-8ecbcc209d02-run-httpd\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.200071 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.206342 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.206783 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-config-data\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.214009 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-scripts\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.226392 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6wg7\" (UniqueName: \"kubernetes.io/projected/80044b30-c554-4654-a4c0-8ecbcc209d02-kube-api-access-v6wg7\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.226473 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gbpf\" (UniqueName: \"kubernetes.io/projected/56ce1365-283f-4f53-8563-60efbce145f7-kube-api-access-2gbpf\") pod \"ironic-db-create-l5q2x\" (UID: \"56ce1365-283f-4f53-8563-60efbce145f7\") " pod="openstack/ironic-db-create-l5q2x" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.234242 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.251954 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-6pgfh"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.269007 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-l5q2x" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.270543 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-8xv6t"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.271802 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.274029 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.274383 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qlg59" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.274549 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.279600 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-vwhfg"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.281274 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.287544 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-8xv6t"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.295547 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8233982e-1252-4dae-a15c-cbdc04652502-db-sync-config-data\") pod \"barbican-db-sync-6pgfh\" (UID: \"8233982e-1252-4dae-a15c-cbdc04652502\") " pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.295723 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8233982e-1252-4dae-a15c-cbdc04652502-combined-ca-bundle\") pod \"barbican-db-sync-6pgfh\" (UID: \"8233982e-1252-4dae-a15c-cbdc04652502\") " pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.295842 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdpzl\" (UniqueName: \"kubernetes.io/projected/8233982e-1252-4dae-a15c-cbdc04652502-kube-api-access-sdpzl\") pod \"barbican-db-sync-6pgfh\" (UID: \"8233982e-1252-4dae-a15c-cbdc04652502\") " pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.300336 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.310767 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-vwhfg"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.329281 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-8924m"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.331007 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.338526 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8924m"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.347803 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.360954 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9jhr2" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.361212 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400206 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5lmr\" (UniqueName: \"kubernetes.io/projected/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-kube-api-access-x5lmr\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400258 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8233982e-1252-4dae-a15c-cbdc04652502-db-sync-config-data\") pod \"barbican-db-sync-6pgfh\" (UID: \"8233982e-1252-4dae-a15c-cbdc04652502\") " pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400287 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-config\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400313 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e71fad98-9eac-426c-9bc5-5d3776a63a8e-logs\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400388 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-scripts\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400412 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400462 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400500 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-config-data\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400537 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400568 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8233982e-1252-4dae-a15c-cbdc04652502-combined-ca-bundle\") pod \"barbican-db-sync-6pgfh\" (UID: \"8233982e-1252-4dae-a15c-cbdc04652502\") " pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400603 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdpzl\" (UniqueName: \"kubernetes.io/projected/8233982e-1252-4dae-a15c-cbdc04652502-kube-api-access-sdpzl\") pod \"barbican-db-sync-6pgfh\" (UID: \"8233982e-1252-4dae-a15c-cbdc04652502\") " pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400630 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-combined-ca-bundle\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.400676 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v49kb\" (UniqueName: \"kubernetes.io/projected/e71fad98-9eac-426c-9bc5-5d3776a63a8e-kube-api-access-v49kb\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.409497 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8233982e-1252-4dae-a15c-cbdc04652502-db-sync-config-data\") pod \"barbican-db-sync-6pgfh\" (UID: \"8233982e-1252-4dae-a15c-cbdc04652502\") " pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.411229 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8233982e-1252-4dae-a15c-cbdc04652502-combined-ca-bundle\") pod \"barbican-db-sync-6pgfh\" (UID: \"8233982e-1252-4dae-a15c-cbdc04652502\") " pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.464554 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdpzl\" (UniqueName: \"kubernetes.io/projected/8233982e-1252-4dae-a15c-cbdc04652502-kube-api-access-sdpzl\") pod \"barbican-db-sync-6pgfh\" (UID: \"8233982e-1252-4dae-a15c-cbdc04652502\") " pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.501893 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-scripts\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.501928 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-combined-ca-bundle\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.501951 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-db-sync-config-data\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.501978 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-combined-ca-bundle\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.501994 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-config-data\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.502027 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v49kb\" (UniqueName: \"kubernetes.io/projected/e71fad98-9eac-426c-9bc5-5d3776a63a8e-kube-api-access-v49kb\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.502052 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5lmr\" (UniqueName: \"kubernetes.io/projected/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-kube-api-access-x5lmr\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.502070 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-config\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.502090 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e71fad98-9eac-426c-9bc5-5d3776a63a8e-logs\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.502116 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-etc-machine-id\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.502140 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk5b2\" (UniqueName: \"kubernetes.io/projected/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-kube-api-access-pk5b2\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.502169 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-scripts\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.502191 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.502212 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.502241 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.502270 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-config-data\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.502297 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.503101 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.504149 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e71fad98-9eac-426c-9bc5-5d3776a63a8e-logs\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.504237 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-config\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.504648 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.504916 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.505418 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.509589 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-combined-ca-bundle\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.509821 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-scripts\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.514076 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-config-data\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.519026 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-rw779"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.520377 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-rw779" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.525208 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-rw779"] Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.533795 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.534081 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-mvgp5" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.534719 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.535053 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v49kb\" (UniqueName: \"kubernetes.io/projected/e71fad98-9eac-426c-9bc5-5d3776a63a8e-kube-api-access-v49kb\") pod \"placement-db-sync-8xv6t\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.560378 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5lmr\" (UniqueName: \"kubernetes.io/projected/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-kube-api-access-x5lmr\") pod \"dnsmasq-dns-fcfdd6f9f-vwhfg\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.598079 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.604733 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-etc-machine-id\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.604793 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk5b2\" (UniqueName: \"kubernetes.io/projected/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-kube-api-access-pk5b2\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.604831 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzgn4\" (UniqueName: \"kubernetes.io/projected/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-kube-api-access-nzgn4\") pod \"neutron-db-sync-rw779\" (UID: \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\") " pod="openstack/neutron-db-sync-rw779" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.604902 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-combined-ca-bundle\") pod \"neutron-db-sync-rw779\" (UID: \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\") " pod="openstack/neutron-db-sync-rw779" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.604978 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-scripts\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.605001 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-combined-ca-bundle\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.605025 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-db-sync-config-data\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.605056 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-config-data\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.605119 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-config\") pod \"neutron-db-sync-rw779\" (UID: \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\") " pod="openstack/neutron-db-sync-rw779" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.605245 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-etc-machine-id\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.607652 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8xv6t" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.609611 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-db-sync-config-data\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.610451 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-config-data\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.610732 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-scripts\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.618950 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.619716 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-combined-ca-bundle\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.626068 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk5b2\" (UniqueName: \"kubernetes.io/projected/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-kube-api-access-pk5b2\") pod \"cinder-db-sync-8924m\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.673513 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8924m" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.706900 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-combined-ca-bundle\") pod \"neutron-db-sync-rw779\" (UID: \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\") " pod="openstack/neutron-db-sync-rw779" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.707079 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-config\") pod \"neutron-db-sync-rw779\" (UID: \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\") " pod="openstack/neutron-db-sync-rw779" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.707170 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzgn4\" (UniqueName: \"kubernetes.io/projected/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-kube-api-access-nzgn4\") pod \"neutron-db-sync-rw779\" (UID: \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\") " pod="openstack/neutron-db-sync-rw779" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.710814 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-combined-ca-bundle\") pod \"neutron-db-sync-rw779\" (UID: \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\") " pod="openstack/neutron-db-sync-rw779" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.714494 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-config\") pod \"neutron-db-sync-rw779\" (UID: \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\") " pod="openstack/neutron-db-sync-rw779" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.726846 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzgn4\" (UniqueName: \"kubernetes.io/projected/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-kube-api-access-nzgn4\") pod \"neutron-db-sync-rw779\" (UID: \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\") " pod="openstack/neutron-db-sync-rw779" Oct 14 15:31:47 crc kubenswrapper[4945]: I1014 15:31:47.857669 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-rw779" Oct 14 15:31:48 crc kubenswrapper[4945]: I1014 15:31:48.544471 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" podUID="9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" containerName="dnsmasq-dns" containerID="cri-o://86b0bcd60848c10e41aff85b5870d6d5aca2d5b1e9d7649a97b2d78f75c903ab" gracePeriod=10 Oct 14 15:31:49 crc kubenswrapper[4945]: I1014 15:31:49.103204 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:31:49 crc kubenswrapper[4945]: I1014 15:31:49.556536 4945 generic.go:334] "Generic (PLEG): container finished" podID="9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" containerID="86b0bcd60848c10e41aff85b5870d6d5aca2d5b1e9d7649a97b2d78f75c903ab" exitCode=0 Oct 14 15:31:49 crc kubenswrapper[4945]: I1014 15:31:49.556582 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" event={"ID":"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62","Type":"ContainerDied","Data":"86b0bcd60848c10e41aff85b5870d6d5aca2d5b1e9d7649a97b2d78f75c903ab"} Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.272094 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.422988 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkb9c\" (UniqueName: \"kubernetes.io/projected/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-kube-api-access-tkb9c\") pod \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.423107 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-nb\") pod \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.423175 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-dns-swift-storage-0\") pod \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.423279 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-sb\") pod \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.423350 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-config\") pod \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.423414 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-dns-svc\") pod \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.441579 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-kube-api-access-tkb9c" (OuterVolumeSpecName: "kube-api-access-tkb9c") pod "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" (UID: "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62"). InnerVolumeSpecName "kube-api-access-tkb9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.474545 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" (UID: "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.475179 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" (UID: "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.498740 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" (UID: "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.503859 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-config" (OuterVolumeSpecName: "config") pod "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" (UID: "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.525707 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" (UID: "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.525820 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-nb\") pod \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\" (UID: \"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62\") " Oct 14 15:31:54 crc kubenswrapper[4945]: W1014 15:31:54.525964 4945 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62/volumes/kubernetes.io~configmap/ovsdbserver-nb Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.525974 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" (UID: "9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.526377 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkb9c\" (UniqueName: \"kubernetes.io/projected/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-kube-api-access-tkb9c\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.526402 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.526411 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.526420 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.526431 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.526440 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.603864 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.607467 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" event={"ID":"9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62","Type":"ContainerDied","Data":"2bbad4e701116d32d38127290bb063676c967e6c22ae4d6406a0548f983f9339"} Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.607517 4945 scope.go:117] "RemoveContainer" containerID="86b0bcd60848c10e41aff85b5870d6d5aca2d5b1e9d7649a97b2d78f75c903ab" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.607613 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.628813 4945 scope.go:117] "RemoveContainer" containerID="a69f5155cd00125a9543770689f463e6b67a7eb4a183c43f70c3393133382750" Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.640116 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-pdrgh"] Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.654718 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-pdrgh"] Oct 14 15:31:54 crc kubenswrapper[4945]: I1014 15:31:54.775680 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" path="/var/lib/kubelet/pods/9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62/volumes" Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.014088 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hgl8j"] Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.067480 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-rw779"] Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.081685 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-8xv6t"] Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.105443 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-create-l5q2x"] Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.140940 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-bjj42"] Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.156288 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-vwhfg"] Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.162192 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8924m"] Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.171665 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-6pgfh"] Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.618824 4945 generic.go:334] "Generic (PLEG): container finished" podID="f6d8cf04-5a11-49aa-ad1d-1129f9974caa" containerID="cd1d1630f43b0367a6890144965ef0127b8d290137f8e36ffae29771e3333a26" exitCode=0 Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.618926 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" event={"ID":"f6d8cf04-5a11-49aa-ad1d-1129f9974caa","Type":"ContainerDied","Data":"cd1d1630f43b0367a6890144965ef0127b8d290137f8e36ffae29771e3333a26"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.618972 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" event={"ID":"f6d8cf04-5a11-49aa-ad1d-1129f9974caa","Type":"ContainerStarted","Data":"6a1dd1b1a49cde3734ad59235945af2942452d037b14d4de3c9010c017c7e85e"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.624095 4945 generic.go:334] "Generic (PLEG): container finished" podID="792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9" containerID="9ba630d026ac4e91fc10ac97c23a60b59992119854ede9dc4ec7ae551836b7cd" exitCode=0 Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.624196 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" event={"ID":"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9","Type":"ContainerDied","Data":"9ba630d026ac4e91fc10ac97c23a60b59992119854ede9dc4ec7ae551836b7cd"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.624252 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" event={"ID":"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9","Type":"ContainerStarted","Data":"5f211bb1f6727797e38c7ef2362d373bc653346a9c6ca6c907c803f4788bb78a"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.632761 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8xv6t" event={"ID":"e71fad98-9eac-426c-9bc5-5d3776a63a8e","Type":"ContainerStarted","Data":"5065a920dc5fa83a330a313f4600deeaa19eacb6ec1ec346e67dcb772b52c28e"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.635358 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mf7rz" event={"ID":"bbc6053c-32ba-4e9f-86b7-8590cff47464","Type":"ContainerStarted","Data":"a70321bd680748812f1478a4e9f2be939302f7d6086028defe66d0c8b4d7c84f"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.650842 4945 generic.go:334] "Generic (PLEG): container finished" podID="56ce1365-283f-4f53-8563-60efbce145f7" containerID="2a8e9691658fc574d29b3377c2db31609ea425cd32d51003ba215c90ed4e78c1" exitCode=0 Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.651089 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-l5q2x" event={"ID":"56ce1365-283f-4f53-8563-60efbce145f7","Type":"ContainerDied","Data":"2a8e9691658fc574d29b3377c2db31609ea425cd32d51003ba215c90ed4e78c1"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.651145 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-l5q2x" event={"ID":"56ce1365-283f-4f53-8563-60efbce145f7","Type":"ContainerStarted","Data":"aed5c204edceb8e172e99e7b8c1f2a9d4d6e7b56c4d78e9a548cd65eac038d4c"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.661279 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hgl8j" event={"ID":"117165e9-bf45-467d-87a8-8237789ca777","Type":"ContainerStarted","Data":"f7bb05eee912710a115a994a9fa6a23574b5055bc7048c2a5c0f429bc98b6492"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.661348 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hgl8j" event={"ID":"117165e9-bf45-467d-87a8-8237789ca777","Type":"ContainerStarted","Data":"aabae1159b13f083499bf8d0bbe530e79aef817372a1d0cbf2725671c63356d0"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.665848 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6pgfh" event={"ID":"8233982e-1252-4dae-a15c-cbdc04652502","Type":"ContainerStarted","Data":"6b282bae5d5faec20f89048f50dad0378b87129371b8472f90b9c383074b7fea"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.669888 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80044b30-c554-4654-a4c0-8ecbcc209d02","Type":"ContainerStarted","Data":"c56d1f4f86a387c8d3bcfcf22d64dca272234525b3ade545fa29fc13a3ec011c"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.673950 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8924m" event={"ID":"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e","Type":"ContainerStarted","Data":"42866bacfe71d8f729ba08606fd19742687b5e83a38d3262a1dd88d6e69719f4"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.682325 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-rw779" event={"ID":"60f0f22a-507b-44f9-97d9-bed2a2d6d43e","Type":"ContainerStarted","Data":"53fc5f52d99e139f5e2baab20bc3a4520e33ed8f4a12f26e927ddf2edcdae561"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.682377 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-rw779" event={"ID":"60f0f22a-507b-44f9-97d9-bed2a2d6d43e","Type":"ContainerStarted","Data":"411a7e59f7af6ee01b37450681f01783d5f6ec4621659b613c30f261e41ce02b"} Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.690651 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-mf7rz" podStartSLOduration=6.741717356 podStartE2EDuration="20.690631397s" podCreationTimestamp="2025-10-14 15:31:35 +0000 UTC" firstStartedPulling="2025-10-14 15:31:40.288035946 +0000 UTC m=+1030.272084314" lastFinishedPulling="2025-10-14 15:31:54.236949977 +0000 UTC m=+1044.220998355" observedRunningTime="2025-10-14 15:31:55.672749455 +0000 UTC m=+1045.656797823" watchObservedRunningTime="2025-10-14 15:31:55.690631397 +0000 UTC m=+1045.674679765" Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.747249 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hgl8j" podStartSLOduration=9.747226946 podStartE2EDuration="9.747226946s" podCreationTimestamp="2025-10-14 15:31:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:31:55.737719054 +0000 UTC m=+1045.721767422" watchObservedRunningTime="2025-10-14 15:31:55.747226946 +0000 UTC m=+1045.731275314" Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.774115 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-rw779" podStartSLOduration=8.774094585 podStartE2EDuration="8.774094585s" podCreationTimestamp="2025-10-14 15:31:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:31:55.753235068 +0000 UTC m=+1045.737283436" watchObservedRunningTime="2025-10-14 15:31:55.774094585 +0000 UTC m=+1045.758142953" Oct 14 15:31:55 crc kubenswrapper[4945]: I1014 15:31:55.959287 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.061310 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-dns-swift-storage-0\") pod \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.061664 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlhgr\" (UniqueName: \"kubernetes.io/projected/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-kube-api-access-hlhgr\") pod \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.061861 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-ovsdbserver-sb\") pod \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.061924 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-config\") pod \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.061948 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-ovsdbserver-nb\") pod \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.062001 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-dns-svc\") pod \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\" (UID: \"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9\") " Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.083140 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-kube-api-access-hlhgr" (OuterVolumeSpecName: "kube-api-access-hlhgr") pod "792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9" (UID: "792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9"). InnerVolumeSpecName "kube-api-access-hlhgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.083811 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9" (UID: "792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.091820 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9" (UID: "792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.100202 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-config" (OuterVolumeSpecName: "config") pod "792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9" (UID: "792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.101823 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9" (UID: "792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.112436 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9" (UID: "792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.174123 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.174455 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlhgr\" (UniqueName: \"kubernetes.io/projected/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-kube-api-access-hlhgr\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.174467 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.174477 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.174489 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.174507 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.695375 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" event={"ID":"f6d8cf04-5a11-49aa-ad1d-1129f9974caa","Type":"ContainerStarted","Data":"800e470818e531c16793af4d95d768b47288c4cea3ca75d7c9df63b585b9eea7"} Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.695530 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.697565 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" event={"ID":"792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9","Type":"ContainerDied","Data":"5f211bb1f6727797e38c7ef2362d373bc653346a9c6ca6c907c803f4788bb78a"} Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.697610 4945 scope.go:117] "RemoveContainer" containerID="9ba630d026ac4e91fc10ac97c23a60b59992119854ede9dc4ec7ae551836b7cd" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.697629 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-bjj42" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.723961 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" podStartSLOduration=9.72394137 podStartE2EDuration="9.72394137s" podCreationTimestamp="2025-10-14 15:31:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:31:56.719158083 +0000 UTC m=+1046.703206451" watchObservedRunningTime="2025-10-14 15:31:56.72394137 +0000 UTC m=+1046.707989738" Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.784526 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-bjj42"] Oct 14 15:31:56 crc kubenswrapper[4945]: I1014 15:31:56.789338 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-bjj42"] Oct 14 15:31:58 crc kubenswrapper[4945]: I1014 15:31:58.774276 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9" path="/var/lib/kubelet/pods/792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9/volumes" Oct 14 15:31:59 crc kubenswrapper[4945]: I1014 15:31:59.090266 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-pdrgh" podUID="9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Oct 14 15:32:02 crc kubenswrapper[4945]: I1014 15:32:02.620088 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:32:02 crc kubenswrapper[4945]: I1014 15:32:02.694663 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-psqrj"] Oct 14 15:32:02 crc kubenswrapper[4945]: I1014 15:32:02.694951 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" podUID="a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" containerName="dnsmasq-dns" containerID="cri-o://d37ceab0048ce9492a641679050a5ad77feb1dcceca706ec58186f08af45e444" gracePeriod=10 Oct 14 15:32:05 crc kubenswrapper[4945]: I1014 15:32:05.762753 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-l5q2x" Oct 14 15:32:05 crc kubenswrapper[4945]: I1014 15:32:05.792388 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-l5q2x" event={"ID":"56ce1365-283f-4f53-8563-60efbce145f7","Type":"ContainerDied","Data":"aed5c204edceb8e172e99e7b8c1f2a9d4d6e7b56c4d78e9a548cd65eac038d4c"} Oct 14 15:32:05 crc kubenswrapper[4945]: I1014 15:32:05.792445 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aed5c204edceb8e172e99e7b8c1f2a9d4d6e7b56c4d78e9a548cd65eac038d4c" Oct 14 15:32:05 crc kubenswrapper[4945]: I1014 15:32:05.792404 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-l5q2x" Oct 14 15:32:05 crc kubenswrapper[4945]: I1014 15:32:05.796180 4945 generic.go:334] "Generic (PLEG): container finished" podID="a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" containerID="d37ceab0048ce9492a641679050a5ad77feb1dcceca706ec58186f08af45e444" exitCode=0 Oct 14 15:32:05 crc kubenswrapper[4945]: I1014 15:32:05.796261 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" event={"ID":"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc","Type":"ContainerDied","Data":"d37ceab0048ce9492a641679050a5ad77feb1dcceca706ec58186f08af45e444"} Oct 14 15:32:05 crc kubenswrapper[4945]: I1014 15:32:05.857509 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gbpf\" (UniqueName: \"kubernetes.io/projected/56ce1365-283f-4f53-8563-60efbce145f7-kube-api-access-2gbpf\") pod \"56ce1365-283f-4f53-8563-60efbce145f7\" (UID: \"56ce1365-283f-4f53-8563-60efbce145f7\") " Oct 14 15:32:05 crc kubenswrapper[4945]: I1014 15:32:05.869135 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56ce1365-283f-4f53-8563-60efbce145f7-kube-api-access-2gbpf" (OuterVolumeSpecName: "kube-api-access-2gbpf") pod "56ce1365-283f-4f53-8563-60efbce145f7" (UID: "56ce1365-283f-4f53-8563-60efbce145f7"). InnerVolumeSpecName "kube-api-access-2gbpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:05 crc kubenswrapper[4945]: I1014 15:32:05.959308 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gbpf\" (UniqueName: \"kubernetes.io/projected/56ce1365-283f-4f53-8563-60efbce145f7-kube-api-access-2gbpf\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:07 crc kubenswrapper[4945]: I1014 15:32:07.210891 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" podUID="a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Oct 14 15:32:09 crc kubenswrapper[4945]: I1014 15:32:09.842677 4945 generic.go:334] "Generic (PLEG): container finished" podID="117165e9-bf45-467d-87a8-8237789ca777" containerID="f7bb05eee912710a115a994a9fa6a23574b5055bc7048c2a5c0f429bc98b6492" exitCode=0 Oct 14 15:32:09 crc kubenswrapper[4945]: I1014 15:32:09.842764 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hgl8j" event={"ID":"117165e9-bf45-467d-87a8-8237789ca777","Type":"ContainerDied","Data":"f7bb05eee912710a115a994a9fa6a23574b5055bc7048c2a5c0f429bc98b6492"} Oct 14 15:32:12 crc kubenswrapper[4945]: I1014 15:32:12.210863 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" podUID="a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Oct 14 15:32:16 crc kubenswrapper[4945]: I1014 15:32:16.796022 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:32:16 crc kubenswrapper[4945]: I1014 15:32:16.796340 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:32:16 crc kubenswrapper[4945]: I1014 15:32:16.796383 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:32:16 crc kubenswrapper[4945]: I1014 15:32:16.797312 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb33e17badcd091cccc29c2b02c75756661752ad2d76d11d53ff1811c318baac"} pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 15:32:16 crc kubenswrapper[4945]: I1014 15:32:16.797404 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" containerID="cri-o://cb33e17badcd091cccc29c2b02c75756661752ad2d76d11d53ff1811c318baac" gracePeriod=600 Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.011960 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-204d-account-create-v7l8c"] Oct 14 15:32:17 crc kubenswrapper[4945]: E1014 15:32:17.012486 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" containerName="dnsmasq-dns" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.012510 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" containerName="dnsmasq-dns" Oct 14 15:32:17 crc kubenswrapper[4945]: E1014 15:32:17.012524 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" containerName="init" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.012533 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" containerName="init" Oct 14 15:32:17 crc kubenswrapper[4945]: E1014 15:32:17.012568 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9" containerName="init" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.012576 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9" containerName="init" Oct 14 15:32:17 crc kubenswrapper[4945]: E1014 15:32:17.012593 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56ce1365-283f-4f53-8563-60efbce145f7" containerName="mariadb-database-create" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.012602 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="56ce1365-283f-4f53-8563-60efbce145f7" containerName="mariadb-database-create" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.012859 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="56ce1365-283f-4f53-8563-60efbce145f7" containerName="mariadb-database-create" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.012924 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="792a72c0-35bc-43a6-aa3c-0be0eeb2ecf9" containerName="init" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.012943 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a1fdf4e-9ca3-4f5d-a5dd-6448afcf7b62" containerName="dnsmasq-dns" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.013661 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-204d-account-create-v7l8c" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.016159 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-db-secret" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.019537 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-204d-account-create-v7l8c"] Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.053907 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lffr2\" (UniqueName: \"kubernetes.io/projected/e6a0ed65-3b9d-460f-9493-4c1280e1b6cc-kube-api-access-lffr2\") pod \"ironic-204d-account-create-v7l8c\" (UID: \"e6a0ed65-3b9d-460f-9493-4c1280e1b6cc\") " pod="openstack/ironic-204d-account-create-v7l8c" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.156173 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lffr2\" (UniqueName: \"kubernetes.io/projected/e6a0ed65-3b9d-460f-9493-4c1280e1b6cc-kube-api-access-lffr2\") pod \"ironic-204d-account-create-v7l8c\" (UID: \"e6a0ed65-3b9d-460f-9493-4c1280e1b6cc\") " pod="openstack/ironic-204d-account-create-v7l8c" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.185523 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lffr2\" (UniqueName: \"kubernetes.io/projected/e6a0ed65-3b9d-460f-9493-4c1280e1b6cc-kube-api-access-lffr2\") pod \"ironic-204d-account-create-v7l8c\" (UID: \"e6a0ed65-3b9d-460f-9493-4c1280e1b6cc\") " pod="openstack/ironic-204d-account-create-v7l8c" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.210328 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" podUID="a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.210468 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.332980 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-204d-account-create-v7l8c" Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.908692 4945 generic.go:334] "Generic (PLEG): container finished" podID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerID="cb33e17badcd091cccc29c2b02c75756661752ad2d76d11d53ff1811c318baac" exitCode=0 Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.908741 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerDied","Data":"cb33e17badcd091cccc29c2b02c75756661752ad2d76d11d53ff1811c318baac"} Oct 14 15:32:17 crc kubenswrapper[4945]: I1014 15:32:17.909896 4945 scope.go:117] "RemoveContainer" containerID="1d2f9a019d15034c004b5841d70bbbf8261549fb60b58e5a8e606324eb14a78f" Oct 14 15:32:18 crc kubenswrapper[4945]: E1014 15:32:18.169150 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 14 15:32:18 crc kubenswrapper[4945]: E1014 15:32:18.169601 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sdpzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-6pgfh_openstack(8233982e-1252-4dae-a15c-cbdc04652502): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:32:18 crc kubenswrapper[4945]: E1014 15:32:18.170922 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-6pgfh" podUID="8233982e-1252-4dae-a15c-cbdc04652502" Oct 14 15:32:18 crc kubenswrapper[4945]: E1014 15:32:18.929539 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-6pgfh" podUID="8233982e-1252-4dae-a15c-cbdc04652502" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.164945 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:32:19 crc kubenswrapper[4945]: E1014 15:32:19.168116 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 14 15:32:19 crc kubenswrapper[4945]: E1014 15:32:19.168264 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pk5b2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-8924m_openstack(ea1a8e32-d4a4-46d5-bedf-5883dbf9761e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:32:19 crc kubenswrapper[4945]: E1014 15:32:19.169344 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-8924m" podUID="ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.180620 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.196309 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-config-data\") pod \"117165e9-bf45-467d-87a8-8237789ca777\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.196381 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-credential-keys\") pod \"117165e9-bf45-467d-87a8-8237789ca777\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.196466 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-scripts\") pod \"117165e9-bf45-467d-87a8-8237789ca777\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.196492 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-fernet-keys\") pod \"117165e9-bf45-467d-87a8-8237789ca777\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.196520 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ngqv\" (UniqueName: \"kubernetes.io/projected/117165e9-bf45-467d-87a8-8237789ca777-kube-api-access-4ngqv\") pod \"117165e9-bf45-467d-87a8-8237789ca777\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.196586 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-combined-ca-bundle\") pod \"117165e9-bf45-467d-87a8-8237789ca777\" (UID: \"117165e9-bf45-467d-87a8-8237789ca777\") " Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.206968 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-scripts" (OuterVolumeSpecName: "scripts") pod "117165e9-bf45-467d-87a8-8237789ca777" (UID: "117165e9-bf45-467d-87a8-8237789ca777"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.216153 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "117165e9-bf45-467d-87a8-8237789ca777" (UID: "117165e9-bf45-467d-87a8-8237789ca777"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.216829 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/117165e9-bf45-467d-87a8-8237789ca777-kube-api-access-4ngqv" (OuterVolumeSpecName: "kube-api-access-4ngqv") pod "117165e9-bf45-467d-87a8-8237789ca777" (UID: "117165e9-bf45-467d-87a8-8237789ca777"). InnerVolumeSpecName "kube-api-access-4ngqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.229965 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "117165e9-bf45-467d-87a8-8237789ca777" (UID: "117165e9-bf45-467d-87a8-8237789ca777"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.256840 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-config-data" (OuterVolumeSpecName: "config-data") pod "117165e9-bf45-467d-87a8-8237789ca777" (UID: "117165e9-bf45-467d-87a8-8237789ca777"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.259205 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "117165e9-bf45-467d-87a8-8237789ca777" (UID: "117165e9-bf45-467d-87a8-8237789ca777"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.298315 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh5vj\" (UniqueName: \"kubernetes.io/projected/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-kube-api-access-mh5vj\") pod \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.298406 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-ovsdbserver-nb\") pod \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.298469 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-config\") pod \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.298592 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-ovsdbserver-sb\") pod \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.298629 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-dns-svc\") pod \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\" (UID: \"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc\") " Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.299039 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.299060 4945 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.299069 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ngqv\" (UniqueName: \"kubernetes.io/projected/117165e9-bf45-467d-87a8-8237789ca777-kube-api-access-4ngqv\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.299077 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.299086 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.299093 4945 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/117165e9-bf45-467d-87a8-8237789ca777-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.302094 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-kube-api-access-mh5vj" (OuterVolumeSpecName: "kube-api-access-mh5vj") pod "a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" (UID: "a9b4d686-da0d-464d-a473-e3e3a1eeb0fc"). InnerVolumeSpecName "kube-api-access-mh5vj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.337788 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" (UID: "a9b4d686-da0d-464d-a473-e3e3a1eeb0fc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.339970 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" (UID: "a9b4d686-da0d-464d-a473-e3e3a1eeb0fc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.342656 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" (UID: "a9b4d686-da0d-464d-a473-e3e3a1eeb0fc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.344762 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-config" (OuterVolumeSpecName: "config") pod "a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" (UID: "a9b4d686-da0d-464d-a473-e3e3a1eeb0fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.400503 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh5vj\" (UniqueName: \"kubernetes.io/projected/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-kube-api-access-mh5vj\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.400536 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.400549 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.400557 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.400565 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.938676 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.938926 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-psqrj" event={"ID":"a9b4d686-da0d-464d-a473-e3e3a1eeb0fc","Type":"ContainerDied","Data":"c15cd82c0982d63df1d6f0326d62087c352b914ed3d8782475eb41fe6b18038f"} Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.944345 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hgl8j" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.945009 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hgl8j" event={"ID":"117165e9-bf45-467d-87a8-8237789ca777","Type":"ContainerDied","Data":"aabae1159b13f083499bf8d0bbe530e79aef817372a1d0cbf2725671c63356d0"} Oct 14 15:32:19 crc kubenswrapper[4945]: E1014 15:32:19.946466 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-8924m" podUID="ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.948324 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aabae1159b13f083499bf8d0bbe530e79aef817372a1d0cbf2725671c63356d0" Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.987246 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-psqrj"] Oct 14 15:32:19 crc kubenswrapper[4945]: I1014 15:32:19.994749 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-psqrj"] Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.255094 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hgl8j"] Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.263559 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hgl8j"] Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.341969 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-kjrhb"] Oct 14 15:32:20 crc kubenswrapper[4945]: E1014 15:32:20.342275 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" containerName="dnsmasq-dns" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.342288 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" containerName="dnsmasq-dns" Oct 14 15:32:20 crc kubenswrapper[4945]: E1014 15:32:20.342304 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="117165e9-bf45-467d-87a8-8237789ca777" containerName="keystone-bootstrap" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.342309 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="117165e9-bf45-467d-87a8-8237789ca777" containerName="keystone-bootstrap" Oct 14 15:32:20 crc kubenswrapper[4945]: E1014 15:32:20.342324 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" containerName="init" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.342331 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" containerName="init" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.342474 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="117165e9-bf45-467d-87a8-8237789ca777" containerName="keystone-bootstrap" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.342485 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" containerName="dnsmasq-dns" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.342997 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.349333 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.349504 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.349715 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-g69v8" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.349922 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.356937 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kjrhb"] Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.417554 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-config-data\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.417659 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-credential-keys\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.417689 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-combined-ca-bundle\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.417829 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-fernet-keys\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.418034 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl5g2\" (UniqueName: \"kubernetes.io/projected/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-kube-api-access-dl5g2\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.418175 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-scripts\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.520036 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-scripts\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.520124 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-config-data\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.520186 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-credential-keys\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.520205 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-combined-ca-bundle\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.520224 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-fernet-keys\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.520255 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl5g2\" (UniqueName: \"kubernetes.io/projected/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-kube-api-access-dl5g2\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.525747 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-config-data\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.527158 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-credential-keys\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.532258 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-scripts\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.536529 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-combined-ca-bundle\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.538081 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-fernet-keys\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.539064 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl5g2\" (UniqueName: \"kubernetes.io/projected/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-kube-api-access-dl5g2\") pod \"keystone-bootstrap-kjrhb\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.681337 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:20 crc kubenswrapper[4945]: E1014 15:32:20.752266 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 14 15:32:20 crc kubenswrapper[4945]: E1014 15:32:20.752435 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v49kb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-8xv6t_openstack(e71fad98-9eac-426c-9bc5-5d3776a63a8e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:32:20 crc kubenswrapper[4945]: E1014 15:32:20.753661 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-8xv6t" podUID="e71fad98-9eac-426c-9bc5-5d3776a63a8e" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.781690 4945 scope.go:117] "RemoveContainer" containerID="d37ceab0048ce9492a641679050a5ad77feb1dcceca706ec58186f08af45e444" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.783507 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="117165e9-bf45-467d-87a8-8237789ca777" path="/var/lib/kubelet/pods/117165e9-bf45-467d-87a8-8237789ca777/volumes" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.784634 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9b4d686-da0d-464d-a473-e3e3a1eeb0fc" path="/var/lib/kubelet/pods/a9b4d686-da0d-464d-a473-e3e3a1eeb0fc/volumes" Oct 14 15:32:20 crc kubenswrapper[4945]: I1014 15:32:20.917452 4945 scope.go:117] "RemoveContainer" containerID="cbcdd1b00aace6e007e092ce8be56bb62ec218c99b74c47ceb71af5b011cbab5" Oct 14 15:32:20 crc kubenswrapper[4945]: E1014 15:32:20.974173 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-8xv6t" podUID="e71fad98-9eac-426c-9bc5-5d3776a63a8e" Oct 14 15:32:21 crc kubenswrapper[4945]: I1014 15:32:21.247692 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-204d-account-create-v7l8c"] Oct 14 15:32:21 crc kubenswrapper[4945]: I1014 15:32:21.308441 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kjrhb"] Oct 14 15:32:21 crc kubenswrapper[4945]: W1014 15:32:21.318249 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13783d41_0f7f_499e_bfbe_eaa0c6440d6e.slice/crio-59b5cee2b116f036208cd6a9dd4fd96c96897548be09ca96ec8a7d50c6e1e40a WatchSource:0}: Error finding container 59b5cee2b116f036208cd6a9dd4fd96c96897548be09ca96ec8a7d50c6e1e40a: Status 404 returned error can't find the container with id 59b5cee2b116f036208cd6a9dd4fd96c96897548be09ca96ec8a7d50c6e1e40a Oct 14 15:32:21 crc kubenswrapper[4945]: I1014 15:32:21.975419 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kjrhb" event={"ID":"13783d41-0f7f-499e-bfbe-eaa0c6440d6e","Type":"ContainerStarted","Data":"8b31e8950fa06614de88fdc5634115683e7ff58dddda2229d1dc4d96c4127b22"} Oct 14 15:32:21 crc kubenswrapper[4945]: I1014 15:32:21.976653 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kjrhb" event={"ID":"13783d41-0f7f-499e-bfbe-eaa0c6440d6e","Type":"ContainerStarted","Data":"59b5cee2b116f036208cd6a9dd4fd96c96897548be09ca96ec8a7d50c6e1e40a"} Oct 14 15:32:21 crc kubenswrapper[4945]: I1014 15:32:21.979335 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80044b30-c554-4654-a4c0-8ecbcc209d02","Type":"ContainerStarted","Data":"e491bfc32b7a037edac29943bbe4cd6cc2899448e598bcb5256c0b05ddcafee4"} Oct 14 15:32:21 crc kubenswrapper[4945]: I1014 15:32:21.982196 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerStarted","Data":"4b33206c225bec396b0367980f7cae9b8b640b794421f9c820e87d051e7c6d3e"} Oct 14 15:32:21 crc kubenswrapper[4945]: I1014 15:32:21.984343 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-204d-account-create-v7l8c" event={"ID":"e6a0ed65-3b9d-460f-9493-4c1280e1b6cc","Type":"ContainerStarted","Data":"26f8c2f34be669431a325b5a51793e9db83dce121f3d64c1f0c493e5f23c7135"} Oct 14 15:32:21 crc kubenswrapper[4945]: I1014 15:32:21.984563 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-204d-account-create-v7l8c" event={"ID":"e6a0ed65-3b9d-460f-9493-4c1280e1b6cc","Type":"ContainerStarted","Data":"b34037cfa38389c433ab23fe721c17e2c5c4b91cedff187de51af1443863b184"} Oct 14 15:32:22 crc kubenswrapper[4945]: I1014 15:32:22.004361 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-kjrhb" podStartSLOduration=2.004342562 podStartE2EDuration="2.004342562s" podCreationTimestamp="2025-10-14 15:32:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:32:21.997319982 +0000 UTC m=+1071.981368350" watchObservedRunningTime="2025-10-14 15:32:22.004342562 +0000 UTC m=+1071.988390930" Oct 14 15:32:22 crc kubenswrapper[4945]: I1014 15:32:22.019297 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-204d-account-create-v7l8c" podStartSLOduration=6.019277449 podStartE2EDuration="6.019277449s" podCreationTimestamp="2025-10-14 15:32:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:32:22.013038761 +0000 UTC m=+1071.997087139" watchObservedRunningTime="2025-10-14 15:32:22.019277449 +0000 UTC m=+1072.003325817" Oct 14 15:32:22 crc kubenswrapper[4945]: I1014 15:32:22.994737 4945 generic.go:334] "Generic (PLEG): container finished" podID="e6a0ed65-3b9d-460f-9493-4c1280e1b6cc" containerID="26f8c2f34be669431a325b5a51793e9db83dce121f3d64c1f0c493e5f23c7135" exitCode=0 Oct 14 15:32:22 crc kubenswrapper[4945]: I1014 15:32:22.994816 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-204d-account-create-v7l8c" event={"ID":"e6a0ed65-3b9d-460f-9493-4c1280e1b6cc","Type":"ContainerDied","Data":"26f8c2f34be669431a325b5a51793e9db83dce121f3d64c1f0c493e5f23c7135"} Oct 14 15:32:24 crc kubenswrapper[4945]: I1014 15:32:24.007477 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80044b30-c554-4654-a4c0-8ecbcc209d02","Type":"ContainerStarted","Data":"fdd15caa5b3c164381c26704f64f97e70d0859940b5c09ef2237003ca120b189"} Oct 14 15:32:24 crc kubenswrapper[4945]: I1014 15:32:24.419366 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-204d-account-create-v7l8c" Oct 14 15:32:24 crc kubenswrapper[4945]: I1014 15:32:24.488859 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lffr2\" (UniqueName: \"kubernetes.io/projected/e6a0ed65-3b9d-460f-9493-4c1280e1b6cc-kube-api-access-lffr2\") pod \"e6a0ed65-3b9d-460f-9493-4c1280e1b6cc\" (UID: \"e6a0ed65-3b9d-460f-9493-4c1280e1b6cc\") " Oct 14 15:32:24 crc kubenswrapper[4945]: I1014 15:32:24.496091 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6a0ed65-3b9d-460f-9493-4c1280e1b6cc-kube-api-access-lffr2" (OuterVolumeSpecName: "kube-api-access-lffr2") pod "e6a0ed65-3b9d-460f-9493-4c1280e1b6cc" (UID: "e6a0ed65-3b9d-460f-9493-4c1280e1b6cc"). InnerVolumeSpecName "kube-api-access-lffr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:24 crc kubenswrapper[4945]: I1014 15:32:24.590762 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lffr2\" (UniqueName: \"kubernetes.io/projected/e6a0ed65-3b9d-460f-9493-4c1280e1b6cc-kube-api-access-lffr2\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:25 crc kubenswrapper[4945]: I1014 15:32:25.023027 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-204d-account-create-v7l8c" Oct 14 15:32:25 crc kubenswrapper[4945]: I1014 15:32:25.023033 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-204d-account-create-v7l8c" event={"ID":"e6a0ed65-3b9d-460f-9493-4c1280e1b6cc","Type":"ContainerDied","Data":"b34037cfa38389c433ab23fe721c17e2c5c4b91cedff187de51af1443863b184"} Oct 14 15:32:25 crc kubenswrapper[4945]: I1014 15:32:25.023083 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b34037cfa38389c433ab23fe721c17e2c5c4b91cedff187de51af1443863b184" Oct 14 15:32:25 crc kubenswrapper[4945]: I1014 15:32:25.026143 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kjrhb" event={"ID":"13783d41-0f7f-499e-bfbe-eaa0c6440d6e","Type":"ContainerDied","Data":"8b31e8950fa06614de88fdc5634115683e7ff58dddda2229d1dc4d96c4127b22"} Oct 14 15:32:25 crc kubenswrapper[4945]: I1014 15:32:25.026088 4945 generic.go:334] "Generic (PLEG): container finished" podID="13783d41-0f7f-499e-bfbe-eaa0c6440d6e" containerID="8b31e8950fa06614de88fdc5634115683e7ff58dddda2229d1dc4d96c4127b22" exitCode=0 Oct 14 15:32:26 crc kubenswrapper[4945]: I1014 15:32:26.048658 4945 generic.go:334] "Generic (PLEG): container finished" podID="bbc6053c-32ba-4e9f-86b7-8590cff47464" containerID="a70321bd680748812f1478a4e9f2be939302f7d6086028defe66d0c8b4d7c84f" exitCode=0 Oct 14 15:32:26 crc kubenswrapper[4945]: I1014 15:32:26.049029 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mf7rz" event={"ID":"bbc6053c-32ba-4e9f-86b7-8590cff47464","Type":"ContainerDied","Data":"a70321bd680748812f1478a4e9f2be939302f7d6086028defe66d0c8b4d7c84f"} Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.243706 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-db-sync-gcqrl"] Oct 14 15:32:27 crc kubenswrapper[4945]: E1014 15:32:27.244867 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6a0ed65-3b9d-460f-9493-4c1280e1b6cc" containerName="mariadb-account-create" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.244905 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6a0ed65-3b9d-460f-9493-4c1280e1b6cc" containerName="mariadb-account-create" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.245213 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6a0ed65-3b9d-460f-9493-4c1280e1b6cc" containerName="mariadb-account-create" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.246544 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.250829 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-config-data" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.251365 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-dockercfg-g7lhl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.255527 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-scripts" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.261214 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-sync-gcqrl"] Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.338368 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/367d71fe-5d2e-470a-958d-4a97ad15545c-etc-podinfo\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.338455 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-config-data\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.338479 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt8br\" (UniqueName: \"kubernetes.io/projected/367d71fe-5d2e-470a-958d-4a97ad15545c-kube-api-access-nt8br\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.338512 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-combined-ca-bundle\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.338529 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-scripts\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.338547 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/367d71fe-5d2e-470a-958d-4a97ad15545c-config-data-merged\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.440691 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/367d71fe-5d2e-470a-958d-4a97ad15545c-etc-podinfo\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.440778 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-config-data\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.440808 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt8br\" (UniqueName: \"kubernetes.io/projected/367d71fe-5d2e-470a-958d-4a97ad15545c-kube-api-access-nt8br\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.440858 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-combined-ca-bundle\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.440901 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-scripts\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.440928 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/367d71fe-5d2e-470a-958d-4a97ad15545c-config-data-merged\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.441572 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/367d71fe-5d2e-470a-958d-4a97ad15545c-config-data-merged\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.451749 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-config-data\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.454338 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-scripts\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.454446 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-combined-ca-bundle\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.463754 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt8br\" (UniqueName: \"kubernetes.io/projected/367d71fe-5d2e-470a-958d-4a97ad15545c-kube-api-access-nt8br\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.472894 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/367d71fe-5d2e-470a-958d-4a97ad15545c-etc-podinfo\") pod \"ironic-db-sync-gcqrl\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:27 crc kubenswrapper[4945]: I1014 15:32:27.576799 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.080842 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mf7rz" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.085826 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:38 crc kubenswrapper[4945]: E1014 15:32:38.143489 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/sg-core:latest" Oct 14 15:32:38 crc kubenswrapper[4945]: E1014 15:32:38.143657 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:sg-core,Image:quay.io/openstack-k8s-operators/sg-core:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sg-core-conf-yaml,ReadOnly:false,MountPath:/etc/sg-core.conf.yaml,SubPath:sg-core.conf.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v6wg7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(80044b30-c554-4654-a4c0-8ecbcc209d02): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.161359 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kjrhb" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.161706 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kjrhb" event={"ID":"13783d41-0f7f-499e-bfbe-eaa0c6440d6e","Type":"ContainerDied","Data":"59b5cee2b116f036208cd6a9dd4fd96c96897548be09ca96ec8a7d50c6e1e40a"} Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.161735 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59b5cee2b116f036208cd6a9dd4fd96c96897548be09ca96ec8a7d50c6e1e40a" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.163113 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mf7rz" event={"ID":"bbc6053c-32ba-4e9f-86b7-8590cff47464","Type":"ContainerDied","Data":"a1526d1ea282ac5bcbeb8c8e0690a0e2a50aaf47b3e6592a83551c6a1a77e683"} Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.163142 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1526d1ea282ac5bcbeb8c8e0690a0e2a50aaf47b3e6592a83551c6a1a77e683" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.163172 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mf7rz" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.249460 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-config-data\") pod \"bbc6053c-32ba-4e9f-86b7-8590cff47464\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.249515 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-fernet-keys\") pod \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.249571 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-config-data\") pod \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.249633 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl5g2\" (UniqueName: \"kubernetes.io/projected/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-kube-api-access-dl5g2\") pod \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.249684 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-combined-ca-bundle\") pod \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.249710 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-scripts\") pod \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.249736 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-combined-ca-bundle\") pod \"bbc6053c-32ba-4e9f-86b7-8590cff47464\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.249759 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-db-sync-config-data\") pod \"bbc6053c-32ba-4e9f-86b7-8590cff47464\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.249791 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-credential-keys\") pod \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\" (UID: \"13783d41-0f7f-499e-bfbe-eaa0c6440d6e\") " Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.249821 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvg67\" (UniqueName: \"kubernetes.io/projected/bbc6053c-32ba-4e9f-86b7-8590cff47464-kube-api-access-rvg67\") pod \"bbc6053c-32ba-4e9f-86b7-8590cff47464\" (UID: \"bbc6053c-32ba-4e9f-86b7-8590cff47464\") " Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.256400 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-scripts" (OuterVolumeSpecName: "scripts") pod "13783d41-0f7f-499e-bfbe-eaa0c6440d6e" (UID: "13783d41-0f7f-499e-bfbe-eaa0c6440d6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.256548 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "13783d41-0f7f-499e-bfbe-eaa0c6440d6e" (UID: "13783d41-0f7f-499e-bfbe-eaa0c6440d6e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.257685 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbc6053c-32ba-4e9f-86b7-8590cff47464-kube-api-access-rvg67" (OuterVolumeSpecName: "kube-api-access-rvg67") pod "bbc6053c-32ba-4e9f-86b7-8590cff47464" (UID: "bbc6053c-32ba-4e9f-86b7-8590cff47464"). InnerVolumeSpecName "kube-api-access-rvg67". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.258652 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "13783d41-0f7f-499e-bfbe-eaa0c6440d6e" (UID: "13783d41-0f7f-499e-bfbe-eaa0c6440d6e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.270568 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bbc6053c-32ba-4e9f-86b7-8590cff47464" (UID: "bbc6053c-32ba-4e9f-86b7-8590cff47464"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.271001 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-kube-api-access-dl5g2" (OuterVolumeSpecName: "kube-api-access-dl5g2") pod "13783d41-0f7f-499e-bfbe-eaa0c6440d6e" (UID: "13783d41-0f7f-499e-bfbe-eaa0c6440d6e"). InnerVolumeSpecName "kube-api-access-dl5g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.278488 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbc6053c-32ba-4e9f-86b7-8590cff47464" (UID: "bbc6053c-32ba-4e9f-86b7-8590cff47464"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.283051 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13783d41-0f7f-499e-bfbe-eaa0c6440d6e" (UID: "13783d41-0f7f-499e-bfbe-eaa0c6440d6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.289064 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-config-data" (OuterVolumeSpecName: "config-data") pod "13783d41-0f7f-499e-bfbe-eaa0c6440d6e" (UID: "13783d41-0f7f-499e-bfbe-eaa0c6440d6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.309034 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-config-data" (OuterVolumeSpecName: "config-data") pod "bbc6053c-32ba-4e9f-86b7-8590cff47464" (UID: "bbc6053c-32ba-4e9f-86b7-8590cff47464"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.354807 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.354850 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.354863 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.354895 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.354908 4945 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.354920 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvg67\" (UniqueName: \"kubernetes.io/projected/bbc6053c-32ba-4e9f-86b7-8590cff47464-kube-api-access-rvg67\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.354933 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc6053c-32ba-4e9f-86b7-8590cff47464-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.354945 4945 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.354955 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:38 crc kubenswrapper[4945]: I1014 15:32:38.354965 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl5g2\" (UniqueName: \"kubernetes.io/projected/13783d41-0f7f-499e-bfbe-eaa0c6440d6e-kube-api-access-dl5g2\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.002060 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-sync-gcqrl"] Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.175041 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-gcqrl" event={"ID":"367d71fe-5d2e-470a-958d-4a97ad15545c","Type":"ContainerStarted","Data":"38e3c0946f017efb5322249021be0e4ab466cde2adcf1d7bc924dda8204aee85"} Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.264586 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5cdf875b4d-nbblw"] Oct 14 15:32:39 crc kubenswrapper[4945]: E1014 15:32:39.265628 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13783d41-0f7f-499e-bfbe-eaa0c6440d6e" containerName="keystone-bootstrap" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.265649 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="13783d41-0f7f-499e-bfbe-eaa0c6440d6e" containerName="keystone-bootstrap" Oct 14 15:32:39 crc kubenswrapper[4945]: E1014 15:32:39.265672 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbc6053c-32ba-4e9f-86b7-8590cff47464" containerName="glance-db-sync" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.265680 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbc6053c-32ba-4e9f-86b7-8590cff47464" containerName="glance-db-sync" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.265866 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbc6053c-32ba-4e9f-86b7-8590cff47464" containerName="glance-db-sync" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.265899 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="13783d41-0f7f-499e-bfbe-eaa0c6440d6e" containerName="keystone-bootstrap" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.266391 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.269768 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.270233 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.270477 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-g69v8" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.271152 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.271327 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.271502 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.278165 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g82r7\" (UniqueName: \"kubernetes.io/projected/fab41439-cbf7-47cd-a439-70519713d3db-kube-api-access-g82r7\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.278211 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-public-tls-certs\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.278257 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-config-data\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.278379 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-combined-ca-bundle\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.278421 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-fernet-keys\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.278446 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-internal-tls-certs\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.278475 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-scripts\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.278508 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-credential-keys\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.281763 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5cdf875b4d-nbblw"] Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.380208 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-combined-ca-bundle\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.380281 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-fernet-keys\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.380304 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-internal-tls-certs\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.380335 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-scripts\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.380374 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-credential-keys\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.380422 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g82r7\" (UniqueName: \"kubernetes.io/projected/fab41439-cbf7-47cd-a439-70519713d3db-kube-api-access-g82r7\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.380447 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-public-tls-certs\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.380485 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-config-data\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.386276 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-config-data\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.391256 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-credential-keys\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.397358 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-scripts\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.397561 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-public-tls-certs\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.414142 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-fernet-keys\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.414654 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-internal-tls-certs\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.422910 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g82r7\" (UniqueName: \"kubernetes.io/projected/fab41439-cbf7-47cd-a439-70519713d3db-kube-api-access-g82r7\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.422942 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fab41439-cbf7-47cd-a439-70519713d3db-combined-ca-bundle\") pod \"keystone-5cdf875b4d-nbblw\" (UID: \"fab41439-cbf7-47cd-a439-70519713d3db\") " pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.503631 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-8jh4x"] Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.519807 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.519959 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-8jh4x"] Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.588667 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.588735 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.588771 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwsb2\" (UniqueName: \"kubernetes.io/projected/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-kube-api-access-dwsb2\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.588834 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-config\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.588860 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.588962 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.624228 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.690926 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-config\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.690966 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.691028 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.691082 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.691107 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.691129 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwsb2\" (UniqueName: \"kubernetes.io/projected/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-kube-api-access-dwsb2\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.692195 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-config\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.692685 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.693233 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.693698 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.694244 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.710758 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwsb2\" (UniqueName: \"kubernetes.io/projected/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-kube-api-access-dwsb2\") pod \"dnsmasq-dns-57c957c4ff-8jh4x\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.846609 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:39 crc kubenswrapper[4945]: I1014 15:32:39.908033 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5cdf875b4d-nbblw"] Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.188438 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8xv6t" event={"ID":"e71fad98-9eac-426c-9bc5-5d3776a63a8e","Type":"ContainerStarted","Data":"778e15952d386fbbecd7166dac417a95d35853e9b3660d8743a32a5cc8604e67"} Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.193528 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8924m" event={"ID":"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e","Type":"ContainerStarted","Data":"86e788c8e24d13173990293bf8dd21fd10234f901f245aba9028807f7451108d"} Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.197527 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5cdf875b4d-nbblw" event={"ID":"fab41439-cbf7-47cd-a439-70519713d3db","Type":"ContainerStarted","Data":"de2073a1794146d2fda80e4d7c03a71c4799b4cb32cc78b56ce8df4b9f4785ea"} Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.206572 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6pgfh" event={"ID":"8233982e-1252-4dae-a15c-cbdc04652502","Type":"ContainerStarted","Data":"de1877f5aac5307e686a80303477a53ed7ad88939c78a0508ec5ac981fe4518b"} Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.218471 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-8xv6t" podStartSLOduration=9.338316767 podStartE2EDuration="53.218450798s" podCreationTimestamp="2025-10-14 15:31:47 +0000 UTC" firstStartedPulling="2025-10-14 15:31:55.081408127 +0000 UTC m=+1045.065456495" lastFinishedPulling="2025-10-14 15:32:38.961542118 +0000 UTC m=+1088.945590526" observedRunningTime="2025-10-14 15:32:40.20107158 +0000 UTC m=+1090.185119949" watchObservedRunningTime="2025-10-14 15:32:40.218450798 +0000 UTC m=+1090.202499166" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.224948 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-8924m" podStartSLOduration=9.362076906 podStartE2EDuration="53.224934563s" podCreationTimestamp="2025-10-14 15:31:47 +0000 UTC" firstStartedPulling="2025-10-14 15:31:55.098356722 +0000 UTC m=+1045.082405110" lastFinishedPulling="2025-10-14 15:32:38.961214359 +0000 UTC m=+1088.945262767" observedRunningTime="2025-10-14 15:32:40.223977336 +0000 UTC m=+1090.208025704" watchObservedRunningTime="2025-10-14 15:32:40.224934563 +0000 UTC m=+1090.208982931" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.243157 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-6pgfh" podStartSLOduration=9.416847193 podStartE2EDuration="53.243129144s" podCreationTimestamp="2025-10-14 15:31:47 +0000 UTC" firstStartedPulling="2025-10-14 15:31:55.133507267 +0000 UTC m=+1045.117555655" lastFinishedPulling="2025-10-14 15:32:38.959789198 +0000 UTC m=+1088.943837606" observedRunningTime="2025-10-14 15:32:40.240305943 +0000 UTC m=+1090.224354321" watchObservedRunningTime="2025-10-14 15:32:40.243129144 +0000 UTC m=+1090.227177522" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.360308 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-8jh4x"] Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.527610 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.529689 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.531592 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-jkrfs" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.531823 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.532293 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.545957 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.606985 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-scripts\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.607544 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-config-data\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.607583 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.607607 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.607630 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc84334-0ba7-4260-bd54-20bad5d8e8b7-logs\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.607667 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adc84334-0ba7-4260-bd54-20bad5d8e8b7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.607694 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8s44\" (UniqueName: \"kubernetes.io/projected/adc84334-0ba7-4260-bd54-20bad5d8e8b7-kube-api-access-j8s44\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.615078 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.616501 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.622847 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.623985 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.709248 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc84334-0ba7-4260-bd54-20bad5d8e8b7-logs\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.709318 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adc84334-0ba7-4260-bd54-20bad5d8e8b7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.709347 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8s44\" (UniqueName: \"kubernetes.io/projected/adc84334-0ba7-4260-bd54-20bad5d8e8b7-kube-api-access-j8s44\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.709375 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-scripts\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.709419 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-config-data\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.709465 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.709489 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.711391 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc84334-0ba7-4260-bd54-20bad5d8e8b7-logs\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.711427 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adc84334-0ba7-4260-bd54-20bad5d8e8b7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.711691 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.717382 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.718497 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-config-data\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.718944 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-scripts\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.738712 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8s44\" (UniqueName: \"kubernetes.io/projected/adc84334-0ba7-4260-bd54-20bad5d8e8b7-kube-api-access-j8s44\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.751527 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.810735 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5af35303-832e-4179-a00e-6c9c3346a87d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.810811 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.810913 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.810991 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5af35303-832e-4179-a00e-6c9c3346a87d-logs\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.811144 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.811183 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4ws4\" (UniqueName: \"kubernetes.io/projected/5af35303-832e-4179-a00e-6c9c3346a87d-kube-api-access-r4ws4\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.811360 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.857916 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.915673 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.915857 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5af35303-832e-4179-a00e-6c9c3346a87d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.915903 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.915930 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.915963 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.915996 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5af35303-832e-4179-a00e-6c9c3346a87d-logs\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.916062 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.916094 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4ws4\" (UniqueName: \"kubernetes.io/projected/5af35303-832e-4179-a00e-6c9c3346a87d-kube-api-access-r4ws4\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.916605 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5af35303-832e-4179-a00e-6c9c3346a87d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.916965 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5af35303-832e-4179-a00e-6c9c3346a87d-logs\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.922433 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.922464 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.924476 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.934837 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4ws4\" (UniqueName: \"kubernetes.io/projected/5af35303-832e-4179-a00e-6c9c3346a87d-kube-api-access-r4ws4\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:40 crc kubenswrapper[4945]: I1014 15:32:40.963162 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:41 crc kubenswrapper[4945]: I1014 15:32:41.231265 4945 generic.go:334] "Generic (PLEG): container finished" podID="b3e29775-9dfb-4b3b-b48a-b0464c5c296d" containerID="a7171f5e54418c36468ddbdd19621bfff73e2b28cfb419d6591a5eddff1c22ba" exitCode=0 Oct 14 15:32:41 crc kubenswrapper[4945]: I1014 15:32:41.231372 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" event={"ID":"b3e29775-9dfb-4b3b-b48a-b0464c5c296d","Type":"ContainerDied","Data":"a7171f5e54418c36468ddbdd19621bfff73e2b28cfb419d6591a5eddff1c22ba"} Oct 14 15:32:41 crc kubenswrapper[4945]: I1014 15:32:41.231425 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" event={"ID":"b3e29775-9dfb-4b3b-b48a-b0464c5c296d","Type":"ContainerStarted","Data":"ed808ca153382d113a6eaa2a3e32bb9ddad57e0eb96182c3ed8274efe5004bce"} Oct 14 15:32:41 crc kubenswrapper[4945]: I1014 15:32:41.236090 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5cdf875b4d-nbblw" event={"ID":"fab41439-cbf7-47cd-a439-70519713d3db","Type":"ContainerStarted","Data":"63f3fe124c02f1839c9e66742ccf7bf22f0d2fb7b0c96b6fd662f6cd9af8b959"} Oct 14 15:32:41 crc kubenswrapper[4945]: I1014 15:32:41.236764 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:32:41 crc kubenswrapper[4945]: I1014 15:32:41.251277 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 15:32:41 crc kubenswrapper[4945]: I1014 15:32:41.278212 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5cdf875b4d-nbblw" podStartSLOduration=2.278189407 podStartE2EDuration="2.278189407s" podCreationTimestamp="2025-10-14 15:32:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:32:41.272188116 +0000 UTC m=+1091.256236484" watchObservedRunningTime="2025-10-14 15:32:41.278189407 +0000 UTC m=+1091.262237775" Oct 14 15:32:41 crc kubenswrapper[4945]: I1014 15:32:41.452566 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:32:41 crc kubenswrapper[4945]: W1014 15:32:41.463728 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadc84334_0ba7_4260_bd54_20bad5d8e8b7.slice/crio-cd72a12e013c9aed30fac6d00fc74a8232b429c32ce56f79c68e6d992044b395 WatchSource:0}: Error finding container cd72a12e013c9aed30fac6d00fc74a8232b429c32ce56f79c68e6d992044b395: Status 404 returned error can't find the container with id cd72a12e013c9aed30fac6d00fc74a8232b429c32ce56f79c68e6d992044b395 Oct 14 15:32:41 crc kubenswrapper[4945]: I1014 15:32:41.926262 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:32:41 crc kubenswrapper[4945]: W1014 15:32:41.927408 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5af35303_832e_4179_a00e_6c9c3346a87d.slice/crio-bf88c62e4acbb340c91559f79b4262367893d1ac0c017d1a28b229ab4d18370f WatchSource:0}: Error finding container bf88c62e4acbb340c91559f79b4262367893d1ac0c017d1a28b229ab4d18370f: Status 404 returned error can't find the container with id bf88c62e4acbb340c91559f79b4262367893d1ac0c017d1a28b229ab4d18370f Oct 14 15:32:42 crc kubenswrapper[4945]: I1014 15:32:42.248805 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adc84334-0ba7-4260-bd54-20bad5d8e8b7","Type":"ContainerStarted","Data":"d6ab6237fb57f889fb964006a306a22142b2a4c647f1fc836619036121255b99"} Oct 14 15:32:42 crc kubenswrapper[4945]: I1014 15:32:42.249193 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adc84334-0ba7-4260-bd54-20bad5d8e8b7","Type":"ContainerStarted","Data":"cd72a12e013c9aed30fac6d00fc74a8232b429c32ce56f79c68e6d992044b395"} Oct 14 15:32:42 crc kubenswrapper[4945]: I1014 15:32:42.251921 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" event={"ID":"b3e29775-9dfb-4b3b-b48a-b0464c5c296d","Type":"ContainerStarted","Data":"bfef77538cfe55588ec4c4d03bbb860a9210f04403ba79e5aac684993b6a79ce"} Oct 14 15:32:42 crc kubenswrapper[4945]: I1014 15:32:42.252084 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:42 crc kubenswrapper[4945]: I1014 15:32:42.256626 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5af35303-832e-4179-a00e-6c9c3346a87d","Type":"ContainerStarted","Data":"bf88c62e4acbb340c91559f79b4262367893d1ac0c017d1a28b229ab4d18370f"} Oct 14 15:32:42 crc kubenswrapper[4945]: I1014 15:32:42.281158 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" podStartSLOduration=3.281137762 podStartE2EDuration="3.281137762s" podCreationTimestamp="2025-10-14 15:32:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:32:42.273310808 +0000 UTC m=+1092.257359186" watchObservedRunningTime="2025-10-14 15:32:42.281137762 +0000 UTC m=+1092.265186140" Oct 14 15:32:42 crc kubenswrapper[4945]: I1014 15:32:42.416340 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:32:42 crc kubenswrapper[4945]: I1014 15:32:42.486693 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:32:43 crc kubenswrapper[4945]: I1014 15:32:43.276453 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5af35303-832e-4179-a00e-6c9c3346a87d","Type":"ContainerStarted","Data":"0ec9ff5df8f203d2f799b52eb146927423852e720b3bfe05be4c33e4a4b9a592"} Oct 14 15:32:44 crc kubenswrapper[4945]: I1014 15:32:44.287779 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5af35303-832e-4179-a00e-6c9c3346a87d","Type":"ContainerStarted","Data":"a8af6b5fb71683b6fef75fe3d3a5f4f79023c41fa4e61dd955d4cb086d00d7db"} Oct 14 15:32:44 crc kubenswrapper[4945]: I1014 15:32:44.290959 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adc84334-0ba7-4260-bd54-20bad5d8e8b7","Type":"ContainerStarted","Data":"960186b0407d5f8464bcdcc6228d6a3a3b0c2846418762cdec4bb21a8fb6e967"} Oct 14 15:32:45 crc kubenswrapper[4945]: I1014 15:32:45.302665 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5af35303-832e-4179-a00e-6c9c3346a87d" containerName="glance-log" containerID="cri-o://0ec9ff5df8f203d2f799b52eb146927423852e720b3bfe05be4c33e4a4b9a592" gracePeriod=30 Oct 14 15:32:45 crc kubenswrapper[4945]: I1014 15:32:45.302763 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5af35303-832e-4179-a00e-6c9c3346a87d" containerName="glance-httpd" containerID="cri-o://a8af6b5fb71683b6fef75fe3d3a5f4f79023c41fa4e61dd955d4cb086d00d7db" gracePeriod=30 Oct 14 15:32:45 crc kubenswrapper[4945]: I1014 15:32:45.303293 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="adc84334-0ba7-4260-bd54-20bad5d8e8b7" containerName="glance-httpd" containerID="cri-o://960186b0407d5f8464bcdcc6228d6a3a3b0c2846418762cdec4bb21a8fb6e967" gracePeriod=30 Oct 14 15:32:45 crc kubenswrapper[4945]: I1014 15:32:45.303297 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="adc84334-0ba7-4260-bd54-20bad5d8e8b7" containerName="glance-log" containerID="cri-o://d6ab6237fb57f889fb964006a306a22142b2a4c647f1fc836619036121255b99" gracePeriod=30 Oct 14 15:32:45 crc kubenswrapper[4945]: I1014 15:32:45.337821 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.3377987430000005 podStartE2EDuration="6.337798743s" podCreationTimestamp="2025-10-14 15:32:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:32:45.336328391 +0000 UTC m=+1095.320376779" watchObservedRunningTime="2025-10-14 15:32:45.337798743 +0000 UTC m=+1095.321847121" Oct 14 15:32:45 crc kubenswrapper[4945]: I1014 15:32:45.372851 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.372830015 podStartE2EDuration="6.372830015s" podCreationTimestamp="2025-10-14 15:32:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:32:45.360159383 +0000 UTC m=+1095.344207751" watchObservedRunningTime="2025-10-14 15:32:45.372830015 +0000 UTC m=+1095.356878403" Oct 14 15:32:46 crc kubenswrapper[4945]: I1014 15:32:46.311007 4945 generic.go:334] "Generic (PLEG): container finished" podID="adc84334-0ba7-4260-bd54-20bad5d8e8b7" containerID="960186b0407d5f8464bcdcc6228d6a3a3b0c2846418762cdec4bb21a8fb6e967" exitCode=0 Oct 14 15:32:46 crc kubenswrapper[4945]: I1014 15:32:46.311037 4945 generic.go:334] "Generic (PLEG): container finished" podID="adc84334-0ba7-4260-bd54-20bad5d8e8b7" containerID="d6ab6237fb57f889fb964006a306a22142b2a4c647f1fc836619036121255b99" exitCode=143 Oct 14 15:32:46 crc kubenswrapper[4945]: I1014 15:32:46.311105 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adc84334-0ba7-4260-bd54-20bad5d8e8b7","Type":"ContainerDied","Data":"960186b0407d5f8464bcdcc6228d6a3a3b0c2846418762cdec4bb21a8fb6e967"} Oct 14 15:32:46 crc kubenswrapper[4945]: I1014 15:32:46.311181 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adc84334-0ba7-4260-bd54-20bad5d8e8b7","Type":"ContainerDied","Data":"d6ab6237fb57f889fb964006a306a22142b2a4c647f1fc836619036121255b99"} Oct 14 15:32:46 crc kubenswrapper[4945]: I1014 15:32:46.312392 4945 generic.go:334] "Generic (PLEG): container finished" podID="5af35303-832e-4179-a00e-6c9c3346a87d" containerID="a8af6b5fb71683b6fef75fe3d3a5f4f79023c41fa4e61dd955d4cb086d00d7db" exitCode=0 Oct 14 15:32:46 crc kubenswrapper[4945]: I1014 15:32:46.312412 4945 generic.go:334] "Generic (PLEG): container finished" podID="5af35303-832e-4179-a00e-6c9c3346a87d" containerID="0ec9ff5df8f203d2f799b52eb146927423852e720b3bfe05be4c33e4a4b9a592" exitCode=143 Oct 14 15:32:46 crc kubenswrapper[4945]: I1014 15:32:46.312430 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5af35303-832e-4179-a00e-6c9c3346a87d","Type":"ContainerDied","Data":"a8af6b5fb71683b6fef75fe3d3a5f4f79023c41fa4e61dd955d4cb086d00d7db"} Oct 14 15:32:46 crc kubenswrapper[4945]: I1014 15:32:46.312456 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5af35303-832e-4179-a00e-6c9c3346a87d","Type":"ContainerDied","Data":"0ec9ff5df8f203d2f799b52eb146927423852e720b3bfe05be4c33e4a4b9a592"} Oct 14 15:32:49 crc kubenswrapper[4945]: I1014 15:32:49.849155 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:32:49 crc kubenswrapper[4945]: I1014 15:32:49.932119 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-vwhfg"] Oct 14 15:32:49 crc kubenswrapper[4945]: I1014 15:32:49.932460 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" podUID="f6d8cf04-5a11-49aa-ad1d-1129f9974caa" containerName="dnsmasq-dns" containerID="cri-o://800e470818e531c16793af4d95d768b47288c4cea3ca75d7c9df63b585b9eea7" gracePeriod=10 Oct 14 15:32:50 crc kubenswrapper[4945]: I1014 15:32:50.358892 4945 generic.go:334] "Generic (PLEG): container finished" podID="f6d8cf04-5a11-49aa-ad1d-1129f9974caa" containerID="800e470818e531c16793af4d95d768b47288c4cea3ca75d7c9df63b585b9eea7" exitCode=0 Oct 14 15:32:50 crc kubenswrapper[4945]: I1014 15:32:50.358947 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" event={"ID":"f6d8cf04-5a11-49aa-ad1d-1129f9974caa","Type":"ContainerDied","Data":"800e470818e531c16793af4d95d768b47288c4cea3ca75d7c9df63b585b9eea7"} Oct 14 15:32:52 crc kubenswrapper[4945]: I1014 15:32:52.379798 4945 generic.go:334] "Generic (PLEG): container finished" podID="e71fad98-9eac-426c-9bc5-5d3776a63a8e" containerID="778e15952d386fbbecd7166dac417a95d35853e9b3660d8743a32a5cc8604e67" exitCode=0 Oct 14 15:32:52 crc kubenswrapper[4945]: I1014 15:32:52.379933 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8xv6t" event={"ID":"e71fad98-9eac-426c-9bc5-5d3776a63a8e","Type":"ContainerDied","Data":"778e15952d386fbbecd7166dac417a95d35853e9b3660d8743a32a5cc8604e67"} Oct 14 15:32:52 crc kubenswrapper[4945]: I1014 15:32:52.619681 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" podUID="f6d8cf04-5a11-49aa-ad1d-1129f9974caa" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: connect: connection refused" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.392767 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5af35303-832e-4179-a00e-6c9c3346a87d","Type":"ContainerDied","Data":"bf88c62e4acbb340c91559f79b4262367893d1ac0c017d1a28b229ab4d18370f"} Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.392818 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf88c62e4acbb340c91559f79b4262367893d1ac0c017d1a28b229ab4d18370f" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.396597 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adc84334-0ba7-4260-bd54-20bad5d8e8b7","Type":"ContainerDied","Data":"cd72a12e013c9aed30fac6d00fc74a8232b429c32ce56f79c68e6d992044b395"} Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.396653 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd72a12e013c9aed30fac6d00fc74a8232b429c32ce56f79c68e6d992044b395" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.434990 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.447313 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551461 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adc84334-0ba7-4260-bd54-20bad5d8e8b7-httpd-run\") pod \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551564 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-combined-ca-bundle\") pod \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551601 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-config-data\") pod \"5af35303-832e-4179-a00e-6c9c3346a87d\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551633 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8s44\" (UniqueName: \"kubernetes.io/projected/adc84334-0ba7-4260-bd54-20bad5d8e8b7-kube-api-access-j8s44\") pod \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551675 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5af35303-832e-4179-a00e-6c9c3346a87d-logs\") pod \"5af35303-832e-4179-a00e-6c9c3346a87d\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551699 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc84334-0ba7-4260-bd54-20bad5d8e8b7-logs\") pod \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551721 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-scripts\") pod \"5af35303-832e-4179-a00e-6c9c3346a87d\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551749 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551791 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-scripts\") pod \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551809 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"5af35303-832e-4179-a00e-6c9c3346a87d\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551886 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-config-data\") pod \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\" (UID: \"adc84334-0ba7-4260-bd54-20bad5d8e8b7\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551938 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-combined-ca-bundle\") pod \"5af35303-832e-4179-a00e-6c9c3346a87d\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551963 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5af35303-832e-4179-a00e-6c9c3346a87d-httpd-run\") pod \"5af35303-832e-4179-a00e-6c9c3346a87d\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.551988 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4ws4\" (UniqueName: \"kubernetes.io/projected/5af35303-832e-4179-a00e-6c9c3346a87d-kube-api-access-r4ws4\") pod \"5af35303-832e-4179-a00e-6c9c3346a87d\" (UID: \"5af35303-832e-4179-a00e-6c9c3346a87d\") " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.553273 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adc84334-0ba7-4260-bd54-20bad5d8e8b7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "adc84334-0ba7-4260-bd54-20bad5d8e8b7" (UID: "adc84334-0ba7-4260-bd54-20bad5d8e8b7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.553673 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5af35303-832e-4179-a00e-6c9c3346a87d-logs" (OuterVolumeSpecName: "logs") pod "5af35303-832e-4179-a00e-6c9c3346a87d" (UID: "5af35303-832e-4179-a00e-6c9c3346a87d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.553770 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adc84334-0ba7-4260-bd54-20bad5d8e8b7-logs" (OuterVolumeSpecName: "logs") pod "adc84334-0ba7-4260-bd54-20bad5d8e8b7" (UID: "adc84334-0ba7-4260-bd54-20bad5d8e8b7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.554057 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5af35303-832e-4179-a00e-6c9c3346a87d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5af35303-832e-4179-a00e-6c9c3346a87d" (UID: "5af35303-832e-4179-a00e-6c9c3346a87d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.560597 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adc84334-0ba7-4260-bd54-20bad5d8e8b7-kube-api-access-j8s44" (OuterVolumeSpecName: "kube-api-access-j8s44") pod "adc84334-0ba7-4260-bd54-20bad5d8e8b7" (UID: "adc84334-0ba7-4260-bd54-20bad5d8e8b7"). InnerVolumeSpecName "kube-api-access-j8s44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.561174 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-scripts" (OuterVolumeSpecName: "scripts") pod "5af35303-832e-4179-a00e-6c9c3346a87d" (UID: "5af35303-832e-4179-a00e-6c9c3346a87d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.562328 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "adc84334-0ba7-4260-bd54-20bad5d8e8b7" (UID: "adc84334-0ba7-4260-bd54-20bad5d8e8b7"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.563032 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "5af35303-832e-4179-a00e-6c9c3346a87d" (UID: "5af35303-832e-4179-a00e-6c9c3346a87d"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.563132 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5af35303-832e-4179-a00e-6c9c3346a87d-kube-api-access-r4ws4" (OuterVolumeSpecName: "kube-api-access-r4ws4") pod "5af35303-832e-4179-a00e-6c9c3346a87d" (UID: "5af35303-832e-4179-a00e-6c9c3346a87d"). InnerVolumeSpecName "kube-api-access-r4ws4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.569092 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-scripts" (OuterVolumeSpecName: "scripts") pod "adc84334-0ba7-4260-bd54-20bad5d8e8b7" (UID: "adc84334-0ba7-4260-bd54-20bad5d8e8b7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.593296 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5af35303-832e-4179-a00e-6c9c3346a87d" (UID: "5af35303-832e-4179-a00e-6c9c3346a87d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.625855 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adc84334-0ba7-4260-bd54-20bad5d8e8b7" (UID: "adc84334-0ba7-4260-bd54-20bad5d8e8b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.627127 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-config-data" (OuterVolumeSpecName: "config-data") pod "5af35303-832e-4179-a00e-6c9c3346a87d" (UID: "5af35303-832e-4179-a00e-6c9c3346a87d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.653928 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.653963 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.653979 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8s44\" (UniqueName: \"kubernetes.io/projected/adc84334-0ba7-4260-bd54-20bad5d8e8b7-kube-api-access-j8s44\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.653994 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5af35303-832e-4179-a00e-6c9c3346a87d-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.654006 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc84334-0ba7-4260-bd54-20bad5d8e8b7-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.654016 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.654054 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.654066 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.654083 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.654095 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af35303-832e-4179-a00e-6c9c3346a87d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.654105 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5af35303-832e-4179-a00e-6c9c3346a87d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.654117 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4ws4\" (UniqueName: \"kubernetes.io/projected/5af35303-832e-4179-a00e-6c9c3346a87d-kube-api-access-r4ws4\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.654127 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adc84334-0ba7-4260-bd54-20bad5d8e8b7-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.673282 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.681729 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-config-data" (OuterVolumeSpecName: "config-data") pod "adc84334-0ba7-4260-bd54-20bad5d8e8b7" (UID: "adc84334-0ba7-4260-bd54-20bad5d8e8b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.684149 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.756419 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.756452 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:53 crc kubenswrapper[4945]: I1014 15:32:53.756462 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc84334-0ba7-4260-bd54-20bad5d8e8b7-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:54 crc kubenswrapper[4945]: E1014 15:32:54.090676 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Oct 14 15:32:54 crc kubenswrapper[4945]: E1014 15:32:54.091127 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v6wg7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(80044b30-c554-4654-a4c0-8ecbcc209d02): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 15:32:54 crc kubenswrapper[4945]: E1014 15:32:54.092291 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"sg-core\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="80044b30-c554-4654-a4c0-8ecbcc209d02" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.407320 4945 generic.go:334] "Generic (PLEG): container finished" podID="8233982e-1252-4dae-a15c-cbdc04652502" containerID="de1877f5aac5307e686a80303477a53ed7ad88939c78a0508ec5ac981fe4518b" exitCode=0 Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.407489 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.407561 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6pgfh" event={"ID":"8233982e-1252-4dae-a15c-cbdc04652502","Type":"ContainerDied","Data":"de1877f5aac5307e686a80303477a53ed7ad88939c78a0508ec5ac981fe4518b"} Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.407627 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="80044b30-c554-4654-a4c0-8ecbcc209d02" containerName="ceilometer-central-agent" containerID="cri-o://e491bfc32b7a037edac29943bbe4cd6cc2899448e598bcb5256c0b05ddcafee4" gracePeriod=30 Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.407703 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.407745 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="80044b30-c554-4654-a4c0-8ecbcc209d02" containerName="ceilometer-notification-agent" containerID="cri-o://fdd15caa5b3c164381c26704f64f97e70d0859940b5c09ef2237003ca120b189" gracePeriod=30 Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.464194 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.473993 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.484499 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.493160 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.516263 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:32:54 crc kubenswrapper[4945]: E1014 15:32:54.516600 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc84334-0ba7-4260-bd54-20bad5d8e8b7" containerName="glance-log" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.516618 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc84334-0ba7-4260-bd54-20bad5d8e8b7" containerName="glance-log" Oct 14 15:32:54 crc kubenswrapper[4945]: E1014 15:32:54.516629 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc84334-0ba7-4260-bd54-20bad5d8e8b7" containerName="glance-httpd" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.516636 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc84334-0ba7-4260-bd54-20bad5d8e8b7" containerName="glance-httpd" Oct 14 15:32:54 crc kubenswrapper[4945]: E1014 15:32:54.516647 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af35303-832e-4179-a00e-6c9c3346a87d" containerName="glance-httpd" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.516653 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af35303-832e-4179-a00e-6c9c3346a87d" containerName="glance-httpd" Oct 14 15:32:54 crc kubenswrapper[4945]: E1014 15:32:54.516673 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af35303-832e-4179-a00e-6c9c3346a87d" containerName="glance-log" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.516679 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af35303-832e-4179-a00e-6c9c3346a87d" containerName="glance-log" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.516910 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af35303-832e-4179-a00e-6c9c3346a87d" containerName="glance-log" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.516933 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af35303-832e-4179-a00e-6c9c3346a87d" containerName="glance-httpd" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.516945 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="adc84334-0ba7-4260-bd54-20bad5d8e8b7" containerName="glance-log" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.516957 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="adc84334-0ba7-4260-bd54-20bad5d8e8b7" containerName="glance-httpd" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.517786 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.520053 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-jkrfs" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.521085 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.521323 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.521488 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.529030 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.530722 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.535028 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.540998 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.541148 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.563856 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569506 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb5c5b1d-4161-435d-8ebe-61016248b180-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569558 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb5c5b1d-4161-435d-8ebe-61016248b180-logs\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569592 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569611 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4rmc\" (UniqueName: \"kubernetes.io/projected/aa12f709-e635-4685-a892-4c65ba588f36-kube-api-access-f4rmc\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569639 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569656 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-config-data\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569674 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa12f709-e635-4685-a892-4c65ba588f36-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569694 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa12f709-e635-4685-a892-4c65ba588f36-logs\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569717 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569736 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4sfg\" (UniqueName: \"kubernetes.io/projected/bb5c5b1d-4161-435d-8ebe-61016248b180-kube-api-access-l4sfg\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569753 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569775 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569794 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-scripts\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569811 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569836 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.569855 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.670768 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa12f709-e635-4685-a892-4c65ba588f36-logs\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.670827 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.670850 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4sfg\" (UniqueName: \"kubernetes.io/projected/bb5c5b1d-4161-435d-8ebe-61016248b180-kube-api-access-l4sfg\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.670885 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.670912 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.670931 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-scripts\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.670951 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.670978 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.670998 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.671034 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb5c5b1d-4161-435d-8ebe-61016248b180-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.671056 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb5c5b1d-4161-435d-8ebe-61016248b180-logs\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.671086 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.671104 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4rmc\" (UniqueName: \"kubernetes.io/projected/aa12f709-e635-4685-a892-4c65ba588f36-kube-api-access-f4rmc\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.671134 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.671151 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-config-data\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.671171 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa12f709-e635-4685-a892-4c65ba588f36-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.671295 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa12f709-e635-4685-a892-4c65ba588f36-logs\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.671569 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa12f709-e635-4685-a892-4c65ba588f36-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.673939 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.674966 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.675242 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb5c5b1d-4161-435d-8ebe-61016248b180-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.680522 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-scripts\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.680652 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.680887 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb5c5b1d-4161-435d-8ebe-61016248b180-logs\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.681341 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.681532 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.682231 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.693860 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.694087 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.695809 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-config-data\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.696863 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4rmc\" (UniqueName: \"kubernetes.io/projected/aa12f709-e635-4685-a892-4c65ba588f36-kube-api-access-f4rmc\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.699249 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4sfg\" (UniqueName: \"kubernetes.io/projected/bb5c5b1d-4161-435d-8ebe-61016248b180-kube-api-access-l4sfg\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.722866 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.726512 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.776576 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5af35303-832e-4179-a00e-6c9c3346a87d" path="/var/lib/kubelet/pods/5af35303-832e-4179-a00e-6c9c3346a87d/volumes" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.777402 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adc84334-0ba7-4260-bd54-20bad5d8e8b7" path="/var/lib/kubelet/pods/adc84334-0ba7-4260-bd54-20bad5d8e8b7/volumes" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.842040 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 15:32:54 crc kubenswrapper[4945]: I1014 15:32:54.858264 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.418265 4945 generic.go:334] "Generic (PLEG): container finished" podID="80044b30-c554-4654-a4c0-8ecbcc209d02" containerID="e491bfc32b7a037edac29943bbe4cd6cc2899448e598bcb5256c0b05ddcafee4" exitCode=0 Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.418395 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80044b30-c554-4654-a4c0-8ecbcc209d02","Type":"ContainerDied","Data":"e491bfc32b7a037edac29943bbe4cd6cc2899448e598bcb5256c0b05ddcafee4"} Oct 14 15:32:55 crc kubenswrapper[4945]: E1014 15:32:55.545171 4945 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified" Oct 14 15:32:55 crc kubenswrapper[4945]: E1014 15:32:55.545393 4945 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/container-scripts/init.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:IronicPassword,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:IronicPassword,Optional:nil,},},},EnvVar{Name:PodName,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:PodNamespace,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:ProvisionNetwork,Value:,ValueFrom:nil,},EnvVar{Name:DatabaseHost,Value:openstack.openstack.svc,ValueFrom:nil,},EnvVar{Name:DatabaseName,Value:ironic,ValueFrom:nil,},EnvVar{Name:DeployHTTPURL,Value:,ValueFrom:nil,},EnvVar{Name:IngressDomain,Value:,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-merged,ReadOnly:false,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-podinfo,ReadOnly:false,MountPath:/etc/podinfo,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nt8br,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-db-sync-gcqrl_openstack(367d71fe-5d2e-470a-958d-4a97ad15545c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 15:32:55 crc kubenswrapper[4945]: E1014 15:32:55.546574 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ironic-db-sync-gcqrl" podUID="367d71fe-5d2e-470a-958d-4a97ad15545c" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.829860 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8xv6t" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.836507 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.845016 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.929513 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-dns-swift-storage-0\") pod \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.929568 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdpzl\" (UniqueName: \"kubernetes.io/projected/8233982e-1252-4dae-a15c-cbdc04652502-kube-api-access-sdpzl\") pod \"8233982e-1252-4dae-a15c-cbdc04652502\" (UID: \"8233982e-1252-4dae-a15c-cbdc04652502\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.929601 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-config-data\") pod \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.929639 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-scripts\") pod \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.929670 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e71fad98-9eac-426c-9bc5-5d3776a63a8e-logs\") pod \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.929691 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-combined-ca-bundle\") pod \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.929718 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-dns-svc\") pod \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.929768 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-ovsdbserver-nb\") pod \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.929976 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v49kb\" (UniqueName: \"kubernetes.io/projected/e71fad98-9eac-426c-9bc5-5d3776a63a8e-kube-api-access-v49kb\") pod \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\" (UID: \"e71fad98-9eac-426c-9bc5-5d3776a63a8e\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.930054 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5lmr\" (UniqueName: \"kubernetes.io/projected/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-kube-api-access-x5lmr\") pod \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.930101 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e71fad98-9eac-426c-9bc5-5d3776a63a8e-logs" (OuterVolumeSpecName: "logs") pod "e71fad98-9eac-426c-9bc5-5d3776a63a8e" (UID: "e71fad98-9eac-426c-9bc5-5d3776a63a8e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.930110 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8233982e-1252-4dae-a15c-cbdc04652502-combined-ca-bundle\") pod \"8233982e-1252-4dae-a15c-cbdc04652502\" (UID: \"8233982e-1252-4dae-a15c-cbdc04652502\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.930181 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-ovsdbserver-sb\") pod \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.930223 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8233982e-1252-4dae-a15c-cbdc04652502-db-sync-config-data\") pod \"8233982e-1252-4dae-a15c-cbdc04652502\" (UID: \"8233982e-1252-4dae-a15c-cbdc04652502\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.930249 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-config\") pod \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\" (UID: \"f6d8cf04-5a11-49aa-ad1d-1129f9974caa\") " Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.930630 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e71fad98-9eac-426c-9bc5-5d3776a63a8e-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.941619 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8233982e-1252-4dae-a15c-cbdc04652502-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8233982e-1252-4dae-a15c-cbdc04652502" (UID: "8233982e-1252-4dae-a15c-cbdc04652502"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.946195 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8233982e-1252-4dae-a15c-cbdc04652502-kube-api-access-sdpzl" (OuterVolumeSpecName: "kube-api-access-sdpzl") pod "8233982e-1252-4dae-a15c-cbdc04652502" (UID: "8233982e-1252-4dae-a15c-cbdc04652502"). InnerVolumeSpecName "kube-api-access-sdpzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.979264 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e71fad98-9eac-426c-9bc5-5d3776a63a8e" (UID: "e71fad98-9eac-426c-9bc5-5d3776a63a8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.980131 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8233982e-1252-4dae-a15c-cbdc04652502-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8233982e-1252-4dae-a15c-cbdc04652502" (UID: "8233982e-1252-4dae-a15c-cbdc04652502"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.981479 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-kube-api-access-x5lmr" (OuterVolumeSpecName: "kube-api-access-x5lmr") pod "f6d8cf04-5a11-49aa-ad1d-1129f9974caa" (UID: "f6d8cf04-5a11-49aa-ad1d-1129f9974caa"). InnerVolumeSpecName "kube-api-access-x5lmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.984073 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f6d8cf04-5a11-49aa-ad1d-1129f9974caa" (UID: "f6d8cf04-5a11-49aa-ad1d-1129f9974caa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.984649 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-scripts" (OuterVolumeSpecName: "scripts") pod "e71fad98-9eac-426c-9bc5-5d3776a63a8e" (UID: "e71fad98-9eac-426c-9bc5-5d3776a63a8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.985749 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e71fad98-9eac-426c-9bc5-5d3776a63a8e-kube-api-access-v49kb" (OuterVolumeSpecName: "kube-api-access-v49kb") pod "e71fad98-9eac-426c-9bc5-5d3776a63a8e" (UID: "e71fad98-9eac-426c-9bc5-5d3776a63a8e"). InnerVolumeSpecName "kube-api-access-v49kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.985999 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-config-data" (OuterVolumeSpecName: "config-data") pod "e71fad98-9eac-426c-9bc5-5d3776a63a8e" (UID: "e71fad98-9eac-426c-9bc5-5d3776a63a8e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.994740 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f6d8cf04-5a11-49aa-ad1d-1129f9974caa" (UID: "f6d8cf04-5a11-49aa-ad1d-1129f9974caa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:55 crc kubenswrapper[4945]: I1014 15:32:55.995922 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f6d8cf04-5a11-49aa-ad1d-1129f9974caa" (UID: "f6d8cf04-5a11-49aa-ad1d-1129f9974caa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.000241 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-config" (OuterVolumeSpecName: "config") pod "f6d8cf04-5a11-49aa-ad1d-1129f9974caa" (UID: "f6d8cf04-5a11-49aa-ad1d-1129f9974caa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.002042 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f6d8cf04-5a11-49aa-ad1d-1129f9974caa" (UID: "f6d8cf04-5a11-49aa-ad1d-1129f9974caa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033100 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033132 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v49kb\" (UniqueName: \"kubernetes.io/projected/e71fad98-9eac-426c-9bc5-5d3776a63a8e-kube-api-access-v49kb\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033144 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5lmr\" (UniqueName: \"kubernetes.io/projected/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-kube-api-access-x5lmr\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033152 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8233982e-1252-4dae-a15c-cbdc04652502-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033160 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033168 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8233982e-1252-4dae-a15c-cbdc04652502-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033177 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033185 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033194 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdpzl\" (UniqueName: \"kubernetes.io/projected/8233982e-1252-4dae-a15c-cbdc04652502-kube-api-access-sdpzl\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033202 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033210 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033217 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71fad98-9eac-426c-9bc5-5d3776a63a8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.033225 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6d8cf04-5a11-49aa-ad1d-1129f9974caa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.129195 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:32:56 crc kubenswrapper[4945]: W1014 15:32:56.151664 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb5c5b1d_4161_435d_8ebe_61016248b180.slice/crio-cb518d21296871c66dfe57407a3a0c0d8cada70748427f63b2d35c4b8cad2a7b WatchSource:0}: Error finding container cb518d21296871c66dfe57407a3a0c0d8cada70748427f63b2d35c4b8cad2a7b: Status 404 returned error can't find the container with id cb518d21296871c66dfe57407a3a0c0d8cada70748427f63b2d35c4b8cad2a7b Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.241286 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:32:56 crc kubenswrapper[4945]: W1014 15:32:56.251243 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa12f709_e635_4685_a892_4c65ba588f36.slice/crio-56c00879666ebf1855e58b9c558528904444d29b6e4a601c5816af8327509e90 WatchSource:0}: Error finding container 56c00879666ebf1855e58b9c558528904444d29b6e4a601c5816af8327509e90: Status 404 returned error can't find the container with id 56c00879666ebf1855e58b9c558528904444d29b6e4a601c5816af8327509e90 Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.430013 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6pgfh" event={"ID":"8233982e-1252-4dae-a15c-cbdc04652502","Type":"ContainerDied","Data":"6b282bae5d5faec20f89048f50dad0378b87129371b8472f90b9c383074b7fea"} Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.430110 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b282bae5d5faec20f89048f50dad0378b87129371b8472f90b9c383074b7fea" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.430257 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6pgfh" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.431776 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aa12f709-e635-4685-a892-4c65ba588f36","Type":"ContainerStarted","Data":"56c00879666ebf1855e58b9c558528904444d29b6e4a601c5816af8327509e90"} Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.442518 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bb5c5b1d-4161-435d-8ebe-61016248b180","Type":"ContainerStarted","Data":"cb518d21296871c66dfe57407a3a0c0d8cada70748427f63b2d35c4b8cad2a7b"} Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.446939 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8xv6t" event={"ID":"e71fad98-9eac-426c-9bc5-5d3776a63a8e","Type":"ContainerDied","Data":"5065a920dc5fa83a330a313f4600deeaa19eacb6ec1ec346e67dcb772b52c28e"} Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.447024 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5065a920dc5fa83a330a313f4600deeaa19eacb6ec1ec346e67dcb772b52c28e" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.447162 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8xv6t" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.453321 4945 generic.go:334] "Generic (PLEG): container finished" podID="ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" containerID="86e788c8e24d13173990293bf8dd21fd10234f901f245aba9028807f7451108d" exitCode=0 Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.453403 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8924m" event={"ID":"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e","Type":"ContainerDied","Data":"86e788c8e24d13173990293bf8dd21fd10234f901f245aba9028807f7451108d"} Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.461364 4945 generic.go:334] "Generic (PLEG): container finished" podID="60f0f22a-507b-44f9-97d9-bed2a2d6d43e" containerID="53fc5f52d99e139f5e2baab20bc3a4520e33ed8f4a12f26e927ddf2edcdae561" exitCode=0 Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.461399 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-rw779" event={"ID":"60f0f22a-507b-44f9-97d9-bed2a2d6d43e","Type":"ContainerDied","Data":"53fc5f52d99e139f5e2baab20bc3a4520e33ed8f4a12f26e927ddf2edcdae561"} Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.466269 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" event={"ID":"f6d8cf04-5a11-49aa-ad1d-1129f9974caa","Type":"ContainerDied","Data":"6a1dd1b1a49cde3734ad59235945af2942452d037b14d4de3c9010c017c7e85e"} Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.466322 4945 scope.go:117] "RemoveContainer" containerID="800e470818e531c16793af4d95d768b47288c4cea3ca75d7c9df63b585b9eea7" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.466603 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-vwhfg" Oct 14 15:32:56 crc kubenswrapper[4945]: E1014 15:32:56.471685 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified\\\"\"" pod="openstack/ironic-db-sync-gcqrl" podUID="367d71fe-5d2e-470a-958d-4a97ad15545c" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.518570 4945 scope.go:117] "RemoveContainer" containerID="cd1d1630f43b0367a6890144965ef0127b8d290137f8e36ffae29771e3333a26" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.538411 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-vwhfg"] Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.545158 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-vwhfg"] Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.756715 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5749bb9b55-8z9wq"] Oct 14 15:32:56 crc kubenswrapper[4945]: E1014 15:32:56.757522 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d8cf04-5a11-49aa-ad1d-1129f9974caa" containerName="dnsmasq-dns" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.757545 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d8cf04-5a11-49aa-ad1d-1129f9974caa" containerName="dnsmasq-dns" Oct 14 15:32:56 crc kubenswrapper[4945]: E1014 15:32:56.757562 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e71fad98-9eac-426c-9bc5-5d3776a63a8e" containerName="placement-db-sync" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.757571 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e71fad98-9eac-426c-9bc5-5d3776a63a8e" containerName="placement-db-sync" Oct 14 15:32:56 crc kubenswrapper[4945]: E1014 15:32:56.757586 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8233982e-1252-4dae-a15c-cbdc04652502" containerName="barbican-db-sync" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.757595 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="8233982e-1252-4dae-a15c-cbdc04652502" containerName="barbican-db-sync" Oct 14 15:32:56 crc kubenswrapper[4945]: E1014 15:32:56.757634 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d8cf04-5a11-49aa-ad1d-1129f9974caa" containerName="init" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.757643 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d8cf04-5a11-49aa-ad1d-1129f9974caa" containerName="init" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.757855 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d8cf04-5a11-49aa-ad1d-1129f9974caa" containerName="dnsmasq-dns" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.757963 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="8233982e-1252-4dae-a15c-cbdc04652502" containerName="barbican-db-sync" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.758005 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e71fad98-9eac-426c-9bc5-5d3776a63a8e" containerName="placement-db-sync" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.759158 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.765572 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-gmlf9" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.765891 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.766099 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.797739 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6d8cf04-5a11-49aa-ad1d-1129f9974caa" path="/var/lib/kubelet/pods/f6d8cf04-5a11-49aa-ad1d-1129f9974caa/volumes" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.811800 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-dfb99d5c8-zqfxx"] Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.814145 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5749bb9b55-8z9wq"] Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.814313 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.814675 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-dfb99d5c8-zqfxx"] Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.816915 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.849916 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-czks5"] Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.850972 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcc68ba8-49a8-456d-b808-2d63dc0d6377-config-data-custom\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.851037 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89w6f\" (UniqueName: \"kubernetes.io/projected/fcc68ba8-49a8-456d-b808-2d63dc0d6377-kube-api-access-89w6f\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.851097 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc68ba8-49a8-456d-b808-2d63dc0d6377-config-data\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.851160 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc68ba8-49a8-456d-b808-2d63dc0d6377-combined-ca-bundle\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.851239 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.851262 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcc68ba8-49a8-456d-b808-2d63dc0d6377-logs\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.902298 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-czks5"] Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.956712 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.956775 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcc68ba8-49a8-456d-b808-2d63dc0d6377-config-data-custom\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.956822 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89w6f\" (UniqueName: \"kubernetes.io/projected/fcc68ba8-49a8-456d-b808-2d63dc0d6377-kube-api-access-89w6f\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.956849 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-config-data-custom\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.963353 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc68ba8-49a8-456d-b808-2d63dc0d6377-config-data\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.963405 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-config\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.963451 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb9jp\" (UniqueName: \"kubernetes.io/projected/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-kube-api-access-bb9jp\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.963478 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-combined-ca-bundle\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.963528 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc68ba8-49a8-456d-b808-2d63dc0d6377-combined-ca-bundle\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.963591 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-logs\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.963648 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.963677 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcc68ba8-49a8-456d-b808-2d63dc0d6377-logs\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.963698 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.963743 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.963800 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-config-data\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.963807 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcc68ba8-49a8-456d-b808-2d63dc0d6377-config-data-custom\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.965496 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-57988bf6fb-7btwh"] Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.964923 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndll2\" (UniqueName: \"kubernetes.io/projected/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-kube-api-access-ndll2\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.967239 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.968383 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcc68ba8-49a8-456d-b808-2d63dc0d6377-logs\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.970941 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.974272 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc68ba8-49a8-456d-b808-2d63dc0d6377-combined-ca-bundle\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.982568 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc68ba8-49a8-456d-b808-2d63dc0d6377-config-data\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.988392 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89w6f\" (UniqueName: \"kubernetes.io/projected/fcc68ba8-49a8-456d-b808-2d63dc0d6377-kube-api-access-89w6f\") pod \"barbican-worker-5749bb9b55-8z9wq\" (UID: \"fcc68ba8-49a8-456d-b808-2d63dc0d6377\") " pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:56 crc kubenswrapper[4945]: I1014 15:32:56.989159 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57988bf6fb-7btwh"] Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.052327 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-bd8df944d-q4z6c"] Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.054071 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.056301 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.056827 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.057019 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.057233 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qlg59" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.057522 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.065412 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bd8df944d-q4z6c"] Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.068164 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-config\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.068501 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb9jp\" (UniqueName: \"kubernetes.io/projected/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-kube-api-access-bb9jp\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.072342 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-config\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.068721 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-combined-ca-bundle\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.074370 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-config-data\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.074448 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfm8n\" (UniqueName: \"kubernetes.io/projected/f93b99df-9b1f-42de-9da1-e537156ec1e9-kube-api-access-pfm8n\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.074547 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-logs\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.074642 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.074702 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.074760 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-combined-ca-bundle\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.074786 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.074949 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-config-data\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.075121 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-logs\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.075576 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-config-data-custom\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.075615 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndll2\" (UniqueName: \"kubernetes.io/projected/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-kube-api-access-ndll2\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.075651 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.075678 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.075742 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-config-data-custom\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.075792 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93b99df-9b1f-42de-9da1-e537156ec1e9-logs\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.076534 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.076539 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.078123 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.078629 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-combined-ca-bundle\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.080251 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-config-data-custom\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.084686 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-config-data\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.093219 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb9jp\" (UniqueName: \"kubernetes.io/projected/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-kube-api-access-bb9jp\") pod \"dnsmasq-dns-6d66f584d7-czks5\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.094484 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndll2\" (UniqueName: \"kubernetes.io/projected/e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67-kube-api-access-ndll2\") pod \"barbican-keystone-listener-dfb99d5c8-zqfxx\" (UID: \"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67\") " pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.104473 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5749bb9b55-8z9wq" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.153194 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.177419 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmfxl\" (UniqueName: \"kubernetes.io/projected/396db799-75e0-4156-a97d-472db3c42729-kube-api-access-bmfxl\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.177743 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/396db799-75e0-4156-a97d-472db3c42729-logs\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.177831 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-config-data-custom\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.177972 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-combined-ca-bundle\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.178108 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-config-data\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.178230 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93b99df-9b1f-42de-9da1-e537156ec1e9-logs\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.178301 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-scripts\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.178352 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-internal-tls-certs\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.178393 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-config-data\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.178423 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfm8n\" (UniqueName: \"kubernetes.io/projected/f93b99df-9b1f-42de-9da1-e537156ec1e9-kube-api-access-pfm8n\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.178687 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-public-tls-certs\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.178737 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-combined-ca-bundle\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.178904 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93b99df-9b1f-42de-9da1-e537156ec1e9-logs\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.186727 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-config-data\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.187660 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-combined-ca-bundle\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.194828 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-config-data-custom\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.199719 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfm8n\" (UniqueName: \"kubernetes.io/projected/f93b99df-9b1f-42de-9da1-e537156ec1e9-kube-api-access-pfm8n\") pod \"barbican-api-57988bf6fb-7btwh\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.280001 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-combined-ca-bundle\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.280414 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-config-data\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.280472 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-scripts\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.280512 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-internal-tls-certs\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.280584 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-public-tls-certs\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.280647 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmfxl\" (UniqueName: \"kubernetes.io/projected/396db799-75e0-4156-a97d-472db3c42729-kube-api-access-bmfxl\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.280674 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/396db799-75e0-4156-a97d-472db3c42729-logs\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.281235 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/396db799-75e0-4156-a97d-472db3c42729-logs\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.288139 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-combined-ca-bundle\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.288971 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-config-data\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.293334 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-internal-tls-certs\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.294400 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-public-tls-certs\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.297175 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/396db799-75e0-4156-a97d-472db3c42729-scripts\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.307198 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmfxl\" (UniqueName: \"kubernetes.io/projected/396db799-75e0-4156-a97d-472db3c42729-kube-api-access-bmfxl\") pod \"placement-bd8df944d-q4z6c\" (UID: \"396db799-75e0-4156-a97d-472db3c42729\") " pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.308818 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.331502 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.378306 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.512855 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aa12f709-e635-4685-a892-4c65ba588f36","Type":"ContainerStarted","Data":"73931af753973ecbc7d6a2fb76ee1d915b3c46b1c26854e339e6a75ae6eb271d"} Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.525796 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bb5c5b1d-4161-435d-8ebe-61016248b180","Type":"ContainerStarted","Data":"166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547"} Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.577826 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5749bb9b55-8z9wq"] Oct 14 15:32:57 crc kubenswrapper[4945]: W1014 15:32:57.642306 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcc68ba8_49a8_456d_b808_2d63dc0d6377.slice/crio-d4f79bc09b9fd97c263230d15dc1c8cbcccf3b9a2dafbd6efa862ec39e04e7eb WatchSource:0}: Error finding container d4f79bc09b9fd97c263230d15dc1c8cbcccf3b9a2dafbd6efa862ec39e04e7eb: Status 404 returned error can't find the container with id d4f79bc09b9fd97c263230d15dc1c8cbcccf3b9a2dafbd6efa862ec39e04e7eb Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.663206 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-dfb99d5c8-zqfxx"] Oct 14 15:32:57 crc kubenswrapper[4945]: I1014 15:32:57.890849 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57988bf6fb-7btwh"] Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.025589 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-czks5"] Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.027791 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bd8df944d-q4z6c"] Oct 14 15:32:58 crc kubenswrapper[4945]: W1014 15:32:58.028284 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode35ba3e3_49ea_49a8_b56f_1aa16c0bfa13.slice/crio-5fd91924c3e56957dc8bb0bcb97c5751540556d276ae4ee27b16a58b41e3ac2f WatchSource:0}: Error finding container 5fd91924c3e56957dc8bb0bcb97c5751540556d276ae4ee27b16a58b41e3ac2f: Status 404 returned error can't find the container with id 5fd91924c3e56957dc8bb0bcb97c5751540556d276ae4ee27b16a58b41e3ac2f Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.088990 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-rw779" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.091354 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8924m" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.207641 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzgn4\" (UniqueName: \"kubernetes.io/projected/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-kube-api-access-nzgn4\") pod \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\" (UID: \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\") " Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.207783 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-combined-ca-bundle\") pod \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.207963 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-etc-machine-id\") pod \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.208186 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-config-data\") pod \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.208305 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk5b2\" (UniqueName: \"kubernetes.io/projected/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-kube-api-access-pk5b2\") pod \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.208421 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-config\") pod \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\" (UID: \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\") " Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.208748 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-scripts\") pod \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.208913 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-combined-ca-bundle\") pod \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\" (UID: \"60f0f22a-507b-44f9-97d9-bed2a2d6d43e\") " Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.209053 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-db-sync-config-data\") pod \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\" (UID: \"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e\") " Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.210712 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" (UID: "ea1a8e32-d4a4-46d5-bedf-5883dbf9761e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.223331 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-kube-api-access-nzgn4" (OuterVolumeSpecName: "kube-api-access-nzgn4") pod "60f0f22a-507b-44f9-97d9-bed2a2d6d43e" (UID: "60f0f22a-507b-44f9-97d9-bed2a2d6d43e"). InnerVolumeSpecName "kube-api-access-nzgn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.226158 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" (UID: "ea1a8e32-d4a4-46d5-bedf-5883dbf9761e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.227124 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-kube-api-access-pk5b2" (OuterVolumeSpecName: "kube-api-access-pk5b2") pod "ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" (UID: "ea1a8e32-d4a4-46d5-bedf-5883dbf9761e"). InnerVolumeSpecName "kube-api-access-pk5b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.229345 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-scripts" (OuterVolumeSpecName: "scripts") pod "ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" (UID: "ea1a8e32-d4a4-46d5-bedf-5883dbf9761e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.282675 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-config" (OuterVolumeSpecName: "config") pod "60f0f22a-507b-44f9-97d9-bed2a2d6d43e" (UID: "60f0f22a-507b-44f9-97d9-bed2a2d6d43e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.283314 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60f0f22a-507b-44f9-97d9-bed2a2d6d43e" (UID: "60f0f22a-507b-44f9-97d9-bed2a2d6d43e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.284471 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" (UID: "ea1a8e32-d4a4-46d5-bedf-5883dbf9761e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.310049 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-config-data" (OuterVolumeSpecName: "config-data") pod "ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" (UID: "ea1a8e32-d4a4-46d5-bedf-5883dbf9761e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.311151 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.311173 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.311185 4945 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.311193 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzgn4\" (UniqueName: \"kubernetes.io/projected/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-kube-api-access-nzgn4\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.311203 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.311211 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.311218 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.311229 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/60f0f22a-507b-44f9-97d9-bed2a2d6d43e-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.311237 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk5b2\" (UniqueName: \"kubernetes.io/projected/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e-kube-api-access-pk5b2\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.535793 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5749bb9b55-8z9wq" event={"ID":"fcc68ba8-49a8-456d-b808-2d63dc0d6377","Type":"ContainerStarted","Data":"d4f79bc09b9fd97c263230d15dc1c8cbcccf3b9a2dafbd6efa862ec39e04e7eb"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.537523 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" event={"ID":"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67","Type":"ContainerStarted","Data":"c00f54e6aaf98e1c169fcac1183fcb896d00f3bb65fd00b035cd966b51cac800"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.539391 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bd8df944d-q4z6c" event={"ID":"396db799-75e0-4156-a97d-472db3c42729","Type":"ContainerStarted","Data":"48951c40024bbb402fce2da5d2bc4d217fbb7c0d4a2afdd3f084cf90d6dffcb8"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.539422 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bd8df944d-q4z6c" event={"ID":"396db799-75e0-4156-a97d-472db3c42729","Type":"ContainerStarted","Data":"20af7075a962089d42a15dd9b867a017b598882b40da04bcef80f7053bb59ec5"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.539434 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bd8df944d-q4z6c" event={"ID":"396db799-75e0-4156-a97d-472db3c42729","Type":"ContainerStarted","Data":"d26ed0faede4fdba48599da38f808855d3b9f41fba02b46a922eabaf315ba7c0"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.541926 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.541969 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.546664 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bb5c5b1d-4161-435d-8ebe-61016248b180","Type":"ContainerStarted","Data":"2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.550187 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57988bf6fb-7btwh" event={"ID":"f93b99df-9b1f-42de-9da1-e537156ec1e9","Type":"ContainerStarted","Data":"b5f7ac0ecd4067d404414c5a558be63bed3ccc4acd14b1957f0c1cf1fae0b94b"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.550233 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57988bf6fb-7btwh" event={"ID":"f93b99df-9b1f-42de-9da1-e537156ec1e9","Type":"ContainerStarted","Data":"7ca2005fe6847cee8addd7898853ce5a01c638da8a163d7854d623dc05eda703"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.550249 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57988bf6fb-7btwh" event={"ID":"f93b99df-9b1f-42de-9da1-e537156ec1e9","Type":"ContainerStarted","Data":"ed160bbb9270ff5eb638c65070f0601723e4ca267f722243004098693a553a7c"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.550412 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.550433 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.552704 4945 generic.go:334] "Generic (PLEG): container finished" podID="e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13" containerID="d09f876f7ba277ae405513414688da99159e23d9ec5d289455192e965c78a405" exitCode=0 Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.552755 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-czks5" event={"ID":"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13","Type":"ContainerDied","Data":"d09f876f7ba277ae405513414688da99159e23d9ec5d289455192e965c78a405"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.552775 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-czks5" event={"ID":"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13","Type":"ContainerStarted","Data":"5fd91924c3e56957dc8bb0bcb97c5751540556d276ae4ee27b16a58b41e3ac2f"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.557428 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8924m" event={"ID":"ea1a8e32-d4a4-46d5-bedf-5883dbf9761e","Type":"ContainerDied","Data":"42866bacfe71d8f729ba08606fd19742687b5e83a38d3262a1dd88d6e69719f4"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.557465 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42866bacfe71d8f729ba08606fd19742687b5e83a38d3262a1dd88d6e69719f4" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.557519 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8924m" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.579538 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-bd8df944d-q4z6c" podStartSLOduration=2.579518749 podStartE2EDuration="2.579518749s" podCreationTimestamp="2025-10-14 15:32:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:32:58.566663471 +0000 UTC m=+1108.550711859" watchObservedRunningTime="2025-10-14 15:32:58.579518749 +0000 UTC m=+1108.563567117" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.581589 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-rw779" event={"ID":"60f0f22a-507b-44f9-97d9-bed2a2d6d43e","Type":"ContainerDied","Data":"411a7e59f7af6ee01b37450681f01783d5f6ec4621659b613c30f261e41ce02b"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.581636 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="411a7e59f7af6ee01b37450681f01783d5f6ec4621659b613c30f261e41ce02b" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.581709 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-rw779" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.613648 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aa12f709-e635-4685-a892-4c65ba588f36","Type":"ContainerStarted","Data":"b241e8b17c3ea5dd6f7019383c566e00ac84c1af7a799b95ea7ad241b3f1a0a8"} Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.744188 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-57988bf6fb-7btwh" podStartSLOduration=2.744167129 podStartE2EDuration="2.744167129s" podCreationTimestamp="2025-10-14 15:32:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:32:58.639576767 +0000 UTC m=+1108.623625145" watchObservedRunningTime="2025-10-14 15:32:58.744167129 +0000 UTC m=+1108.728215497" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.791588 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.791570125 podStartE2EDuration="4.791570125s" podCreationTimestamp="2025-10-14 15:32:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:32:58.684479672 +0000 UTC m=+1108.668528050" watchObservedRunningTime="2025-10-14 15:32:58.791570125 +0000 UTC m=+1108.775618493" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.857860 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 15:32:58 crc kubenswrapper[4945]: E1014 15:32:58.858283 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" containerName="cinder-db-sync" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.858299 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" containerName="cinder-db-sync" Oct 14 15:32:58 crc kubenswrapper[4945]: E1014 15:32:58.858319 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60f0f22a-507b-44f9-97d9-bed2a2d6d43e" containerName="neutron-db-sync" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.858325 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="60f0f22a-507b-44f9-97d9-bed2a2d6d43e" containerName="neutron-db-sync" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.858505 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" containerName="cinder-db-sync" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.858519 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="60f0f22a-507b-44f9-97d9-bed2a2d6d43e" containerName="neutron-db-sync" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.859426 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.861767 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9jhr2" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.862162 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.862451 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.862694 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-czks5"] Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.862792 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.877317 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.885154 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-l5fl8"] Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.890305 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.892110 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-794d4664d6-ns65b"] Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.894948 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.898276 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-l5fl8"] Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.900519 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.900762 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.901008 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.901371 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-mvgp5" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.921071 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-794d4664d6-ns65b"] Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.942045 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.94202261 podStartE2EDuration="4.94202261s" podCreationTimestamp="2025-10-14 15:32:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:32:58.802541239 +0000 UTC m=+1108.786589607" watchObservedRunningTime="2025-10-14 15:32:58.94202261 +0000 UTC m=+1108.926070978" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.970969 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-l5fl8"] Oct 14 15:32:58 crc kubenswrapper[4945]: E1014 15:32:58.971608 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-n6rhb ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" podUID="d86a77f4-156c-4941-8252-f32bc9157b8e" Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.977363 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-49n26"] Oct 14 15:32:58 crc kubenswrapper[4945]: I1014 15:32:58.979050 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.043592 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-49n26"] Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045607 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045653 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-ovndb-tls-certs\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045677 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045701 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4fcl\" (UniqueName: \"kubernetes.io/projected/786749d9-049d-4391-8220-8eb69462f5d1-kube-api-access-p4fcl\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045723 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045744 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfsmr\" (UniqueName: \"kubernetes.io/projected/82646027-6477-4a53-bbcc-cdf19a9e03da-kube-api-access-bfsmr\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045770 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-config\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045800 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-combined-ca-bundle\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045822 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-scripts\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045841 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045866 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82646027-6477-4a53-bbcc-cdf19a9e03da-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045910 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045931 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-config\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045955 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-dns-svc\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.045998 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-config-data\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.046021 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-httpd-config\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.046048 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6rhb\" (UniqueName: \"kubernetes.io/projected/d86a77f4-156c-4941-8252-f32bc9157b8e-kube-api-access-n6rhb\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.110933 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.113723 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.136999 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.137012 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148263 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148339 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-ovndb-tls-certs\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148405 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148459 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4fcl\" (UniqueName: \"kubernetes.io/projected/786749d9-049d-4391-8220-8eb69462f5d1-kube-api-access-p4fcl\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148498 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148534 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfsmr\" (UniqueName: \"kubernetes.io/projected/82646027-6477-4a53-bbcc-cdf19a9e03da-kube-api-access-bfsmr\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148574 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-config\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148606 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148638 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-config\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148701 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148742 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-combined-ca-bundle\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148775 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-scripts\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148814 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148853 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82646027-6477-4a53-bbcc-cdf19a9e03da-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148922 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148952 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdjc8\" (UniqueName: \"kubernetes.io/projected/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-kube-api-access-mdjc8\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.148984 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-config\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.149020 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-dns-svc\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.149117 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.149150 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-config-data\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.149199 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-httpd-config\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.149302 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6rhb\" (UniqueName: \"kubernetes.io/projected/d86a77f4-156c-4941-8252-f32bc9157b8e-kube-api-access-n6rhb\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.149409 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.152939 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-config\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.155551 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.155647 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.157012 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-dns-svc\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.157708 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-config-data\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.157733 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.161992 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82646027-6477-4a53-bbcc-cdf19a9e03da-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.164607 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.171686 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-httpd-config\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.172502 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-config\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.173282 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-scripts\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.190515 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6rhb\" (UniqueName: \"kubernetes.io/projected/d86a77f4-156c-4941-8252-f32bc9157b8e-kube-api-access-n6rhb\") pod \"dnsmasq-dns-688c87cc99-l5fl8\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.193815 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfsmr\" (UniqueName: \"kubernetes.io/projected/82646027-6477-4a53-bbcc-cdf19a9e03da-kube-api-access-bfsmr\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.202618 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-ovndb-tls-certs\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.203157 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-combined-ca-bundle\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.203421 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4fcl\" (UniqueName: \"kubernetes.io/projected/786749d9-049d-4391-8220-8eb69462f5d1-kube-api-access-p4fcl\") pod \"neutron-794d4664d6-ns65b\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.204577 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.254701 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255247 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-scripts\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255309 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-config-data-custom\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255368 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ntqp\" (UniqueName: \"kubernetes.io/projected/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-kube-api-access-4ntqp\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255399 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-config\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255427 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255483 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255505 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255548 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-logs\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255568 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-config-data\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255780 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdjc8\" (UniqueName: \"kubernetes.io/projected/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-kube-api-access-mdjc8\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255846 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.255905 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.256631 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.256665 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-config\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.257420 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.258217 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.258610 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: E1014 15:32:59.266829 4945 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 14 15:32:59 crc kubenswrapper[4945]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 14 15:32:59 crc kubenswrapper[4945]: > podSandboxID="5fd91924c3e56957dc8bb0bcb97c5751540556d276ae4ee27b16a58b41e3ac2f" Oct 14 15:32:59 crc kubenswrapper[4945]: E1014 15:32:59.266987 4945 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 14 15:32:59 crc kubenswrapper[4945]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n554h646h555h654hd6h55fh5ch576h55dh5b5h697h6dh55bh699h56h569hd6hb5h549h675h66h57dh64bhbh96h7fh76hc8h547h7bh699h5cbq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bb9jp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6d66f584d7-czks5_openstack(e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 14 15:32:59 crc kubenswrapper[4945]: > logger="UnhandledError" Oct 14 15:32:59 crc kubenswrapper[4945]: E1014 15:32:59.269547 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-6d66f584d7-czks5" podUID="e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.286556 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdjc8\" (UniqueName: \"kubernetes.io/projected/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-kube-api-access-mdjc8\") pod \"dnsmasq-dns-6bb4fc677f-49n26\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.358254 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-logs\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.358594 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-config-data\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.358899 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-scripts\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.358932 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-config-data-custom\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.359006 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ntqp\" (UniqueName: \"kubernetes.io/projected/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-kube-api-access-4ntqp\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.359046 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.359069 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.359165 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.360256 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-logs\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.365869 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.365943 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-config-data-custom\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.369481 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-config-data\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.370105 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-scripts\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.377593 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ntqp\" (UniqueName: \"kubernetes.io/projected/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-kube-api-access-4ntqp\") pod \"cinder-api-0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.414592 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.465565 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.495808 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.643828 4945 generic.go:334] "Generic (PLEG): container finished" podID="80044b30-c554-4654-a4c0-8ecbcc209d02" containerID="fdd15caa5b3c164381c26704f64f97e70d0859940b5c09ef2237003ca120b189" exitCode=0 Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.644110 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80044b30-c554-4654-a4c0-8ecbcc209d02","Type":"ContainerDied","Data":"fdd15caa5b3c164381c26704f64f97e70d0859940b5c09ef2237003ca120b189"} Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.644179 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.711437 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.769536 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-ovsdbserver-nb\") pod \"d86a77f4-156c-4941-8252-f32bc9157b8e\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.769806 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6rhb\" (UniqueName: \"kubernetes.io/projected/d86a77f4-156c-4941-8252-f32bc9157b8e-kube-api-access-n6rhb\") pod \"d86a77f4-156c-4941-8252-f32bc9157b8e\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.769857 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-ovsdbserver-sb\") pod \"d86a77f4-156c-4941-8252-f32bc9157b8e\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.769926 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-dns-swift-storage-0\") pod \"d86a77f4-156c-4941-8252-f32bc9157b8e\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.770022 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-dns-svc\") pod \"d86a77f4-156c-4941-8252-f32bc9157b8e\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.770068 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-config\") pod \"d86a77f4-156c-4941-8252-f32bc9157b8e\" (UID: \"d86a77f4-156c-4941-8252-f32bc9157b8e\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.770522 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d86a77f4-156c-4941-8252-f32bc9157b8e" (UID: "d86a77f4-156c-4941-8252-f32bc9157b8e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.770781 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d86a77f4-156c-4941-8252-f32bc9157b8e" (UID: "d86a77f4-156c-4941-8252-f32bc9157b8e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.770866 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d86a77f4-156c-4941-8252-f32bc9157b8e" (UID: "d86a77f4-156c-4941-8252-f32bc9157b8e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.771463 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d86a77f4-156c-4941-8252-f32bc9157b8e" (UID: "d86a77f4-156c-4941-8252-f32bc9157b8e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.771764 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-config" (OuterVolumeSpecName: "config") pod "d86a77f4-156c-4941-8252-f32bc9157b8e" (UID: "d86a77f4-156c-4941-8252-f32bc9157b8e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.775672 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d86a77f4-156c-4941-8252-f32bc9157b8e-kube-api-access-n6rhb" (OuterVolumeSpecName: "kube-api-access-n6rhb") pod "d86a77f4-156c-4941-8252-f32bc9157b8e" (UID: "d86a77f4-156c-4941-8252-f32bc9157b8e"). InnerVolumeSpecName "kube-api-access-n6rhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.831593 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.872890 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.872937 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.872948 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.872964 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6rhb\" (UniqueName: \"kubernetes.io/projected/d86a77f4-156c-4941-8252-f32bc9157b8e-kube-api-access-n6rhb\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.872976 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.872987 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d86a77f4-156c-4941-8252-f32bc9157b8e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.947966 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-49n26"] Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.974713 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-sg-core-conf-yaml\") pod \"80044b30-c554-4654-a4c0-8ecbcc209d02\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.974811 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6wg7\" (UniqueName: \"kubernetes.io/projected/80044b30-c554-4654-a4c0-8ecbcc209d02-kube-api-access-v6wg7\") pod \"80044b30-c554-4654-a4c0-8ecbcc209d02\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.974841 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80044b30-c554-4654-a4c0-8ecbcc209d02-log-httpd\") pod \"80044b30-c554-4654-a4c0-8ecbcc209d02\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.975039 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-config-data\") pod \"80044b30-c554-4654-a4c0-8ecbcc209d02\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.975079 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-combined-ca-bundle\") pod \"80044b30-c554-4654-a4c0-8ecbcc209d02\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.975124 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-scripts\") pod \"80044b30-c554-4654-a4c0-8ecbcc209d02\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.975142 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80044b30-c554-4654-a4c0-8ecbcc209d02-run-httpd\") pod \"80044b30-c554-4654-a4c0-8ecbcc209d02\" (UID: \"80044b30-c554-4654-a4c0-8ecbcc209d02\") " Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.975956 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80044b30-c554-4654-a4c0-8ecbcc209d02-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "80044b30-c554-4654-a4c0-8ecbcc209d02" (UID: "80044b30-c554-4654-a4c0-8ecbcc209d02"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.979157 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80044b30-c554-4654-a4c0-8ecbcc209d02-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "80044b30-c554-4654-a4c0-8ecbcc209d02" (UID: "80044b30-c554-4654-a4c0-8ecbcc209d02"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.981647 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "80044b30-c554-4654-a4c0-8ecbcc209d02" (UID: "80044b30-c554-4654-a4c0-8ecbcc209d02"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.983815 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80044b30-c554-4654-a4c0-8ecbcc209d02-kube-api-access-v6wg7" (OuterVolumeSpecName: "kube-api-access-v6wg7") pod "80044b30-c554-4654-a4c0-8ecbcc209d02" (UID: "80044b30-c554-4654-a4c0-8ecbcc209d02"). InnerVolumeSpecName "kube-api-access-v6wg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:32:59 crc kubenswrapper[4945]: I1014 15:32:59.987179 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-scripts" (OuterVolumeSpecName: "scripts") pod "80044b30-c554-4654-a4c0-8ecbcc209d02" (UID: "80044b30-c554-4654-a4c0-8ecbcc209d02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.010123 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-794d4664d6-ns65b"] Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.036786 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.048599 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80044b30-c554-4654-a4c0-8ecbcc209d02" (UID: "80044b30-c554-4654-a4c0-8ecbcc209d02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.049988 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-config-data" (OuterVolumeSpecName: "config-data") pod "80044b30-c554-4654-a4c0-8ecbcc209d02" (UID: "80044b30-c554-4654-a4c0-8ecbcc209d02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.077583 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.077615 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6wg7\" (UniqueName: \"kubernetes.io/projected/80044b30-c554-4654-a4c0-8ecbcc209d02-kube-api-access-v6wg7\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.077629 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80044b30-c554-4654-a4c0-8ecbcc209d02-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.077637 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.077646 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.077654 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80044b30-c554-4654-a4c0-8ecbcc209d02-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.077662 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80044b30-c554-4654-a4c0-8ecbcc209d02-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.150070 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 15:33:00 crc kubenswrapper[4945]: W1014 15:33:00.268415 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82646027_6477_4a53_bbcc_cdf19a9e03da.slice/crio-024b4748b9703b10aa1f05cd033ba09d61b3a10d6227661b49539bba76a6a9e4 WatchSource:0}: Error finding container 024b4748b9703b10aa1f05cd033ba09d61b3a10d6227661b49539bba76a6a9e4: Status 404 returned error can't find the container with id 024b4748b9703b10aa1f05cd033ba09d61b3a10d6227661b49539bba76a6a9e4 Oct 14 15:33:00 crc kubenswrapper[4945]: W1014 15:33:00.270739 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d3bb3a8_c4b8_41b6_b61e_f99f8eb48be0.slice/crio-77717072fda9cc1869cc3d438cc66e53218e9e4e5261ed73ed79c876912c5af0 WatchSource:0}: Error finding container 77717072fda9cc1869cc3d438cc66e53218e9e4e5261ed73ed79c876912c5af0: Status 404 returned error can't find the container with id 77717072fda9cc1869cc3d438cc66e53218e9e4e5261ed73ed79c876912c5af0 Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.283491 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.381596 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-config\") pod \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.381962 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-ovsdbserver-nb\") pod \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.382089 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb9jp\" (UniqueName: \"kubernetes.io/projected/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-kube-api-access-bb9jp\") pod \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.382148 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-dns-swift-storage-0\") pod \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.382190 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-ovsdbserver-sb\") pod \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.382267 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-dns-svc\") pod \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\" (UID: \"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13\") " Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.385660 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-kube-api-access-bb9jp" (OuterVolumeSpecName: "kube-api-access-bb9jp") pod "e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13" (UID: "e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13"). InnerVolumeSpecName "kube-api-access-bb9jp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.434814 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13" (UID: "e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.435492 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13" (UID: "e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.437235 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-config" (OuterVolumeSpecName: "config") pod "e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13" (UID: "e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.440773 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13" (UID: "e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.441332 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13" (UID: "e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.483911 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.483951 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.483962 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.483971 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.483981 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb9jp\" (UniqueName: \"kubernetes.io/projected/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-kube-api-access-bb9jp\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.483992 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:00 crc kubenswrapper[4945]: W1014 15:33:00.610620 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod786749d9_049d_4391_8220_8eb69462f5d1.slice/crio-baf7d5fc59e2fb444f9738369733286209c618b8b4d512f631199eae47f201a7 WatchSource:0}: Error finding container baf7d5fc59e2fb444f9738369733286209c618b8b4d512f631199eae47f201a7: Status 404 returned error can't find the container with id baf7d5fc59e2fb444f9738369733286209c618b8b4d512f631199eae47f201a7 Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.654529 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" event={"ID":"ea523baf-11e5-412e-b5ea-84b3dff1bcb3","Type":"ContainerStarted","Data":"0807b194ee7d42fc4abd8bc8cf071e1a6fe3697edc5ca0ca6590095fd58b72e9"} Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.656253 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0","Type":"ContainerStarted","Data":"77717072fda9cc1869cc3d438cc66e53218e9e4e5261ed73ed79c876912c5af0"} Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.657507 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-czks5" event={"ID":"e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13","Type":"ContainerDied","Data":"5fd91924c3e56957dc8bb0bcb97c5751540556d276ae4ee27b16a58b41e3ac2f"} Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.657540 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-czks5" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.657548 4945 scope.go:117] "RemoveContainer" containerID="d09f876f7ba277ae405513414688da99159e23d9ec5d289455192e965c78a405" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.673998 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"82646027-6477-4a53-bbcc-cdf19a9e03da","Type":"ContainerStarted","Data":"024b4748b9703b10aa1f05cd033ba09d61b3a10d6227661b49539bba76a6a9e4"} Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.679037 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80044b30-c554-4654-a4c0-8ecbcc209d02","Type":"ContainerDied","Data":"c56d1f4f86a387c8d3bcfcf22d64dca272234525b3ade545fa29fc13a3ec011c"} Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.679283 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.685813 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-794d4664d6-ns65b" event={"ID":"786749d9-049d-4391-8220-8eb69462f5d1","Type":"ContainerStarted","Data":"baf7d5fc59e2fb444f9738369733286209c618b8b4d512f631199eae47f201a7"} Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.685882 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-l5fl8" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.814271 4945 scope.go:117] "RemoveContainer" containerID="fdd15caa5b3c164381c26704f64f97e70d0859940b5c09ef2237003ca120b189" Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.949681 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-l5fl8"] Oct 14 15:33:00 crc kubenswrapper[4945]: I1014 15:33:00.972402 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-l5fl8"] Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.015987 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.040952 4945 scope.go:117] "RemoveContainer" containerID="e491bfc32b7a037edac29943bbe4cd6cc2899448e598bcb5256c0b05ddcafee4" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.041095 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.081643 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:01 crc kubenswrapper[4945]: E1014 15:33:01.083616 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13" containerName="init" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.083639 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13" containerName="init" Oct 14 15:33:01 crc kubenswrapper[4945]: E1014 15:33:01.083677 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80044b30-c554-4654-a4c0-8ecbcc209d02" containerName="ceilometer-central-agent" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.083684 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="80044b30-c554-4654-a4c0-8ecbcc209d02" containerName="ceilometer-central-agent" Oct 14 15:33:01 crc kubenswrapper[4945]: E1014 15:33:01.083703 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80044b30-c554-4654-a4c0-8ecbcc209d02" containerName="ceilometer-notification-agent" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.083711 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="80044b30-c554-4654-a4c0-8ecbcc209d02" containerName="ceilometer-notification-agent" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.084152 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="80044b30-c554-4654-a4c0-8ecbcc209d02" containerName="ceilometer-central-agent" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.084174 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="80044b30-c554-4654-a4c0-8ecbcc209d02" containerName="ceilometer-notification-agent" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.084198 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13" containerName="init" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.086265 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.088767 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.090229 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.095003 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-czks5"] Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.115664 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-czks5"] Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.124867 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.211460 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-config-data\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.211511 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/706a95cb-0318-4cad-af7a-04324bb19b23-log-httpd\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.211588 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-scripts\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.211696 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.211721 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.211783 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/706a95cb-0318-4cad-af7a-04324bb19b23-run-httpd\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.211827 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsm4h\" (UniqueName: \"kubernetes.io/projected/706a95cb-0318-4cad-af7a-04324bb19b23-kube-api-access-dsm4h\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.313225 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/706a95cb-0318-4cad-af7a-04324bb19b23-run-httpd\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.313306 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsm4h\" (UniqueName: \"kubernetes.io/projected/706a95cb-0318-4cad-af7a-04324bb19b23-kube-api-access-dsm4h\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.313370 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-config-data\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.313401 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/706a95cb-0318-4cad-af7a-04324bb19b23-log-httpd\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.313459 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-scripts\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.313585 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.313635 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.315421 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/706a95cb-0318-4cad-af7a-04324bb19b23-log-httpd\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.315421 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/706a95cb-0318-4cad-af7a-04324bb19b23-run-httpd\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.319833 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.320393 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-config-data\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.321387 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-scripts\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.331718 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.346030 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsm4h\" (UniqueName: \"kubernetes.io/projected/706a95cb-0318-4cad-af7a-04324bb19b23-kube-api-access-dsm4h\") pod \"ceilometer-0\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.420794 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.761769 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" event={"ID":"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67","Type":"ContainerStarted","Data":"e6a51c64111753cc8543e897ba0bf32f9c14df7c5de786cd85d29e519f136872"} Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.762242 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" event={"ID":"e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67","Type":"ContainerStarted","Data":"0645ec76806aff81e86be9abb89f40560ce0bbcef85546a2eb473732651788c8"} Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.772999 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-794d4664d6-ns65b" event={"ID":"786749d9-049d-4391-8220-8eb69462f5d1","Type":"ContainerStarted","Data":"d70f25ff17a42303a4c62305888fa4fe15b5e474589473276093b7e05c215022"} Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.773070 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-794d4664d6-ns65b" event={"ID":"786749d9-049d-4391-8220-8eb69462f5d1","Type":"ContainerStarted","Data":"a73b648136829109d27da9b0af403d40da63e0076e21a11215b81e2a1c1b0fd4"} Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.773323 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.775350 4945 generic.go:334] "Generic (PLEG): container finished" podID="ea523baf-11e5-412e-b5ea-84b3dff1bcb3" containerID="b31e359b3fa23de12774859bfd90254efd89dace829bc5094b2504daa003f084" exitCode=0 Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.778252 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" event={"ID":"ea523baf-11e5-412e-b5ea-84b3dff1bcb3","Type":"ContainerDied","Data":"b31e359b3fa23de12774859bfd90254efd89dace829bc5094b2504daa003f084"} Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.799417 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0","Type":"ContainerStarted","Data":"05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85"} Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.808067 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-dfb99d5c8-zqfxx" podStartSLOduration=2.809147288 podStartE2EDuration="5.808042527s" podCreationTimestamp="2025-10-14 15:32:56 +0000 UTC" firstStartedPulling="2025-10-14 15:32:57.686228172 +0000 UTC m=+1107.670276540" lastFinishedPulling="2025-10-14 15:33:00.685123411 +0000 UTC m=+1110.669171779" observedRunningTime="2025-10-14 15:33:01.778617665 +0000 UTC m=+1111.762666053" watchObservedRunningTime="2025-10-14 15:33:01.808042527 +0000 UTC m=+1111.792090895" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.817952 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5749bb9b55-8z9wq" event={"ID":"fcc68ba8-49a8-456d-b808-2d63dc0d6377","Type":"ContainerStarted","Data":"74f728f3c79d9b2d5332cf7ad75526c1bc9894a435a5440996aa96611f7b0d84"} Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.817992 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5749bb9b55-8z9wq" event={"ID":"fcc68ba8-49a8-456d-b808-2d63dc0d6377","Type":"ContainerStarted","Data":"f572a15bfe0355d6e9ae9f21a81cb74c0c34f91ddebda9dd737e821f21e3466b"} Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.851306 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-794d4664d6-ns65b" podStartSLOduration=3.851284414 podStartE2EDuration="3.851284414s" podCreationTimestamp="2025-10-14 15:32:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:33:01.83786853 +0000 UTC m=+1111.821916898" watchObservedRunningTime="2025-10-14 15:33:01.851284414 +0000 UTC m=+1111.835332782" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.923465 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5749bb9b55-8z9wq" podStartSLOduration=2.8868378310000002 podStartE2EDuration="5.923448438s" podCreationTimestamp="2025-10-14 15:32:56 +0000 UTC" firstStartedPulling="2025-10-14 15:32:57.649478971 +0000 UTC m=+1107.633527339" lastFinishedPulling="2025-10-14 15:33:00.686089568 +0000 UTC m=+1110.670137946" observedRunningTime="2025-10-14 15:33:01.871946405 +0000 UTC m=+1111.855994773" watchObservedRunningTime="2025-10-14 15:33:01.923448438 +0000 UTC m=+1111.907496806" Oct 14 15:33:01 crc kubenswrapper[4945]: I1014 15:33:01.928893 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:02 crc kubenswrapper[4945]: I1014 15:33:02.780462 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80044b30-c554-4654-a4c0-8ecbcc209d02" path="/var/lib/kubelet/pods/80044b30-c554-4654-a4c0-8ecbcc209d02/volumes" Oct 14 15:33:02 crc kubenswrapper[4945]: I1014 15:33:02.781813 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d86a77f4-156c-4941-8252-f32bc9157b8e" path="/var/lib/kubelet/pods/d86a77f4-156c-4941-8252-f32bc9157b8e/volumes" Oct 14 15:33:02 crc kubenswrapper[4945]: I1014 15:33:02.782266 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13" path="/var/lib/kubelet/pods/e35ba3e3-49ea-49a8-b56f-1aa16c0bfa13/volumes" Oct 14 15:33:02 crc kubenswrapper[4945]: I1014 15:33:02.833408 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" event={"ID":"ea523baf-11e5-412e-b5ea-84b3dff1bcb3","Type":"ContainerStarted","Data":"0e69d3d567d012f6328607ddd7ea353000e2d899addbff317eed50910ffc2ad6"} Oct 14 15:33:02 crc kubenswrapper[4945]: I1014 15:33:02.833770 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:33:02 crc kubenswrapper[4945]: I1014 15:33:02.835710 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0","Type":"ContainerStarted","Data":"7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84"} Oct 14 15:33:02 crc kubenswrapper[4945]: I1014 15:33:02.836277 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 14 15:33:02 crc kubenswrapper[4945]: I1014 15:33:02.838115 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"82646027-6477-4a53-bbcc-cdf19a9e03da","Type":"ContainerStarted","Data":"f9269ae02ee5b157b988e20f7d2940b824ee467a8bc14eef9bf251081291b6be"} Oct 14 15:33:02 crc kubenswrapper[4945]: I1014 15:33:02.842023 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"706a95cb-0318-4cad-af7a-04324bb19b23","Type":"ContainerStarted","Data":"300fdb03106ab359adb493f73ea9942c25a4297c01ea32eb9d39f657633ba37e"} Oct 14 15:33:02 crc kubenswrapper[4945]: I1014 15:33:02.858032 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" podStartSLOduration=4.858010137 podStartE2EDuration="4.858010137s" podCreationTimestamp="2025-10-14 15:32:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:33:02.853571899 +0000 UTC m=+1112.837620267" watchObservedRunningTime="2025-10-14 15:33:02.858010137 +0000 UTC m=+1112.842058495" Oct 14 15:33:02 crc kubenswrapper[4945]: I1014 15:33:02.888209 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.88818447 podStartE2EDuration="4.88818447s" podCreationTimestamp="2025-10-14 15:32:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:33:02.88153566 +0000 UTC m=+1112.865584038" watchObservedRunningTime="2025-10-14 15:33:02.88818447 +0000 UTC m=+1112.872232838" Oct 14 15:33:03 crc kubenswrapper[4945]: I1014 15:33:03.279369 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 15:33:03 crc kubenswrapper[4945]: I1014 15:33:03.870303 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"82646027-6477-4a53-bbcc-cdf19a9e03da","Type":"ContainerStarted","Data":"1dea1f38fddd12ad1d406319300028dfceaa3449363ead51e289edeacc699c94"} Oct 14 15:33:03 crc kubenswrapper[4945]: I1014 15:33:03.872420 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"706a95cb-0318-4cad-af7a-04324bb19b23","Type":"ContainerStarted","Data":"ddbba1cf3884c0569f9d8f731c0f13f71e0f040a7b2e85d865178c53f8d2b484"} Oct 14 15:33:03 crc kubenswrapper[4945]: I1014 15:33:03.898014 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.071237587 podStartE2EDuration="5.897999091s" podCreationTimestamp="2025-10-14 15:32:58 +0000 UTC" firstStartedPulling="2025-10-14 15:33:00.27196504 +0000 UTC m=+1110.256013398" lastFinishedPulling="2025-10-14 15:33:01.098726534 +0000 UTC m=+1111.082774902" observedRunningTime="2025-10-14 15:33:03.893574114 +0000 UTC m=+1113.877622482" watchObservedRunningTime="2025-10-14 15:33:03.897999091 +0000 UTC m=+1113.882047459" Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.497421 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.842950 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.843013 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.859283 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.859328 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.886863 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"706a95cb-0318-4cad-af7a-04324bb19b23","Type":"ContainerStarted","Data":"aa819694cf6724c3d6eda56f7d1dcda666b4d83a95af317fef6782410af75377"} Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.887089 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" containerName="cinder-api-log" containerID="cri-o://05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85" gracePeriod=30 Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.887188 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" containerName="cinder-api" containerID="cri-o://7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84" gracePeriod=30 Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.898340 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.898759 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.909549 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.914733 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.916010 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:04 crc kubenswrapper[4945]: I1014 15:33:04.957128 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.585732 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.623815 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-combined-ca-bundle\") pod \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.623961 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ntqp\" (UniqueName: \"kubernetes.io/projected/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-kube-api-access-4ntqp\") pod \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.623989 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-etc-machine-id\") pod \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.624033 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-logs\") pod \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.624064 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-config-data-custom\") pod \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.624083 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-scripts\") pod \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.624103 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-config-data\") pod \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\" (UID: \"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0\") " Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.624960 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" (UID: "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.625389 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-logs" (OuterVolumeSpecName: "logs") pod "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" (UID: "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.631276 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-scripts" (OuterVolumeSpecName: "scripts") pod "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" (UID: "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.633058 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-kube-api-access-4ntqp" (OuterVolumeSpecName: "kube-api-access-4ntqp") pod "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" (UID: "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0"). InnerVolumeSpecName "kube-api-access-4ntqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.633720 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" (UID: "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.671049 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" (UID: "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.677795 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-config-data" (OuterVolumeSpecName: "config-data") pod "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" (UID: "0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.725685 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.725720 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ntqp\" (UniqueName: \"kubernetes.io/projected/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-kube-api-access-4ntqp\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.725736 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.725748 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.725763 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.725773 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.725784 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.897363 4945 generic.go:334] "Generic (PLEG): container finished" podID="0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" containerID="7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84" exitCode=0 Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.897420 4945 generic.go:334] "Generic (PLEG): container finished" podID="0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" containerID="05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85" exitCode=143 Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.897419 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.897504 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0","Type":"ContainerDied","Data":"7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84"} Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.897569 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0","Type":"ContainerDied","Data":"05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85"} Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.897585 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0","Type":"ContainerDied","Data":"77717072fda9cc1869cc3d438cc66e53218e9e4e5261ed73ed79c876912c5af0"} Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.897603 4945 scope.go:117] "RemoveContainer" containerID="7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.901598 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"706a95cb-0318-4cad-af7a-04324bb19b23","Type":"ContainerStarted","Data":"e466723f709f1ff3739636b9b6ee0f4ccaf3c5a203968a48cf2c38df46d2777b"} Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.901732 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.902225 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.933868 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.940109 4945 scope.go:117] "RemoveContainer" containerID="05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.944100 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.955804 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 15:33:05 crc kubenswrapper[4945]: E1014 15:33:05.956718 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" containerName="cinder-api" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.956737 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" containerName="cinder-api" Oct 14 15:33:05 crc kubenswrapper[4945]: E1014 15:33:05.956755 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" containerName="cinder-api-log" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.956764 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" containerName="cinder-api-log" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.956933 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" containerName="cinder-api-log" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.956959 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" containerName="cinder-api" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.957950 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.961635 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.961727 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.961829 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.971652 4945 scope.go:117] "RemoveContainer" containerID="7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84" Oct 14 15:33:05 crc kubenswrapper[4945]: E1014 15:33:05.972119 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84\": container with ID starting with 7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84 not found: ID does not exist" containerID="7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.972154 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84"} err="failed to get container status \"7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84\": rpc error: code = NotFound desc = could not find container \"7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84\": container with ID starting with 7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84 not found: ID does not exist" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.972182 4945 scope.go:117] "RemoveContainer" containerID="05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85" Oct 14 15:33:05 crc kubenswrapper[4945]: E1014 15:33:05.972461 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85\": container with ID starting with 05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85 not found: ID does not exist" containerID="05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.972519 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85"} err="failed to get container status \"05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85\": rpc error: code = NotFound desc = could not find container \"05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85\": container with ID starting with 05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85 not found: ID does not exist" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.972551 4945 scope.go:117] "RemoveContainer" containerID="7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.973383 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84"} err="failed to get container status \"7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84\": rpc error: code = NotFound desc = could not find container \"7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84\": container with ID starting with 7a49bbaf14e3dc3f0ac65015ed07c08addca56c0901cb68e4fd81f3286359d84 not found: ID does not exist" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.973412 4945 scope.go:117] "RemoveContainer" containerID="05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85" Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.973484 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 15:33:05 crc kubenswrapper[4945]: I1014 15:33:05.973819 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85"} err="failed to get container status \"05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85\": rpc error: code = NotFound desc = could not find container \"05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85\": container with ID starting with 05efc2171a7d3c17132d9eda3bdddce532712447d6655b48b46da4a3b88bdd85 not found: ID does not exist" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.032920 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.033006 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b210a0b1-468b-4e67-8618-02cb27c78033-logs\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.033062 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.033105 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-scripts\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.033143 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-config-data-custom\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.033176 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-config-data\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.033210 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b210a0b1-468b-4e67-8618-02cb27c78033-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.033237 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n24dt\" (UniqueName: \"kubernetes.io/projected/b210a0b1-468b-4e67-8618-02cb27c78033-kube-api-access-n24dt\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.033299 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.097318 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7cb8f5594f-b9dt5"] Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.098855 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.102312 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.102737 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134303 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-public-tls-certs\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134390 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134426 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-config\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134462 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-scripts\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134496 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-config-data-custom\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134522 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-config-data\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134551 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b210a0b1-468b-4e67-8618-02cb27c78033-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134578 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n24dt\" (UniqueName: \"kubernetes.io/projected/b210a0b1-468b-4e67-8618-02cb27c78033-kube-api-access-n24dt\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134634 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134670 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-combined-ca-bundle\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134734 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfjxp\" (UniqueName: \"kubernetes.io/projected/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-kube-api-access-jfjxp\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134826 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134853 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-ovndb-tls-certs\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134880 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-httpd-config\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.134958 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-internal-tls-certs\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.135130 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b210a0b1-468b-4e67-8618-02cb27c78033-logs\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.135700 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b210a0b1-468b-4e67-8618-02cb27c78033-logs\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.136625 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b210a0b1-468b-4e67-8618-02cb27c78033-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.137927 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7cb8f5594f-b9dt5"] Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.141626 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-config-data-custom\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.142411 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-config-data\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.142849 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.143978 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.147554 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.170523 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b210a0b1-468b-4e67-8618-02cb27c78033-scripts\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.179394 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n24dt\" (UniqueName: \"kubernetes.io/projected/b210a0b1-468b-4e67-8618-02cb27c78033-kube-api-access-n24dt\") pod \"cinder-api-0\" (UID: \"b210a0b1-468b-4e67-8618-02cb27c78033\") " pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.236964 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-public-tls-certs\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.237931 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-config\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.238042 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-combined-ca-bundle\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.238081 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfjxp\" (UniqueName: \"kubernetes.io/projected/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-kube-api-access-jfjxp\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.238100 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-ovndb-tls-certs\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.238144 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-httpd-config\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.238162 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-internal-tls-certs\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.242036 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-public-tls-certs\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.247387 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-httpd-config\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.247950 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-combined-ca-bundle\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.248484 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-internal-tls-certs\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.257523 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-ovndb-tls-certs\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.258109 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-config\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.271566 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfjxp\" (UniqueName: \"kubernetes.io/projected/904476f1-4b6d-432f-9db6-a0fd60f6ac3a-kube-api-access-jfjxp\") pod \"neutron-7cb8f5594f-b9dt5\" (UID: \"904476f1-4b6d-432f-9db6-a0fd60f6ac3a\") " pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.279663 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.432728 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.781083 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0" path="/var/lib/kubelet/pods/0d3bb3a8-c4b8-41b6-b61e-f99f8eb48be0/volumes" Oct 14 15:33:06 crc kubenswrapper[4945]: W1014 15:33:06.835461 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb210a0b1_468b_4e67_8618_02cb27c78033.slice/crio-ee79d81122cf52abb89b306ece2323631e7cae9857008aa89897ab70b3e68975 WatchSource:0}: Error finding container ee79d81122cf52abb89b306ece2323631e7cae9857008aa89897ab70b3e68975: Status 404 returned error can't find the container with id ee79d81122cf52abb89b306ece2323631e7cae9857008aa89897ab70b3e68975 Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.839018 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.926762 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b210a0b1-468b-4e67-8618-02cb27c78033","Type":"ContainerStarted","Data":"ee79d81122cf52abb89b306ece2323631e7cae9857008aa89897ab70b3e68975"} Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.938004 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"706a95cb-0318-4cad-af7a-04324bb19b23","Type":"ContainerStarted","Data":"8a3426b80153b65b4a2094de34d3d5b9ae3026558616871dfe038a68769a0ca7"} Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.939073 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.943950 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 15:33:06 crc kubenswrapper[4945]: I1014 15:33:06.944558 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 15:33:07 crc kubenswrapper[4945]: I1014 15:33:07.115367 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.513428178 podStartE2EDuration="7.115346589s" podCreationTimestamp="2025-10-14 15:33:00 +0000 UTC" firstStartedPulling="2025-10-14 15:33:01.960762256 +0000 UTC m=+1111.944810624" lastFinishedPulling="2025-10-14 15:33:06.562680667 +0000 UTC m=+1116.546729035" observedRunningTime="2025-10-14 15:33:06.972852352 +0000 UTC m=+1116.956900720" watchObservedRunningTime="2025-10-14 15:33:07.115346589 +0000 UTC m=+1117.099394957" Oct 14 15:33:07 crc kubenswrapper[4945]: W1014 15:33:07.127037 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod904476f1_4b6d_432f_9db6_a0fd60f6ac3a.slice/crio-7fdaf3b9de9405ae0f457d3b9d82330b161b5550ab6dff71c77f328fd4e500c5 WatchSource:0}: Error finding container 7fdaf3b9de9405ae0f457d3b9d82330b161b5550ab6dff71c77f328fd4e500c5: Status 404 returned error can't find the container with id 7fdaf3b9de9405ae0f457d3b9d82330b161b5550ab6dff71c77f328fd4e500c5 Oct 14 15:33:07 crc kubenswrapper[4945]: I1014 15:33:07.127130 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7cb8f5594f-b9dt5"] Oct 14 15:33:07 crc kubenswrapper[4945]: I1014 15:33:07.468839 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 15:33:07 crc kubenswrapper[4945]: I1014 15:33:07.469513 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 15:33:07 crc kubenswrapper[4945]: I1014 15:33:07.595735 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:07 crc kubenswrapper[4945]: I1014 15:33:07.595789 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:07 crc kubenswrapper[4945]: I1014 15:33:07.983855 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b210a0b1-468b-4e67-8618-02cb27c78033","Type":"ContainerStarted","Data":"f20bf22f95a8b1e2944ad1f638af6ac04386b63d8f7510f54cd9a7baf6ab1c0b"} Oct 14 15:33:07 crc kubenswrapper[4945]: I1014 15:33:07.987609 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cb8f5594f-b9dt5" event={"ID":"904476f1-4b6d-432f-9db6-a0fd60f6ac3a","Type":"ContainerStarted","Data":"9299b7dfc8a455fe931e17fbe12f0f5f568f12166fde9992cd5f3cecb61af6bc"} Oct 14 15:33:07 crc kubenswrapper[4945]: I1014 15:33:07.987640 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cb8f5594f-b9dt5" event={"ID":"904476f1-4b6d-432f-9db6-a0fd60f6ac3a","Type":"ContainerStarted","Data":"24000d67f847cfa74d7bbd174992f3cf3e8a7fb3f501cb2046049bcd9e5a43a2"} Oct 14 15:33:07 crc kubenswrapper[4945]: I1014 15:33:07.987652 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cb8f5594f-b9dt5" event={"ID":"904476f1-4b6d-432f-9db6-a0fd60f6ac3a","Type":"ContainerStarted","Data":"7fdaf3b9de9405ae0f457d3b9d82330b161b5550ab6dff71c77f328fd4e500c5"} Oct 14 15:33:07 crc kubenswrapper[4945]: I1014 15:33:07.992006 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7f6f97bcf4-qbb2j"] Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:07.993557 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.011773 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.016807 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.041992 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f6f97bcf4-qbb2j"] Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.098481 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7cb8f5594f-b9dt5" podStartSLOduration=2.098463396 podStartE2EDuration="2.098463396s" podCreationTimestamp="2025-10-14 15:33:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:33:08.012986851 +0000 UTC m=+1117.997035239" watchObservedRunningTime="2025-10-14 15:33:08.098463396 +0000 UTC m=+1118.082511764" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.123278 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-internal-tls-certs\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.123395 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-public-tls-certs\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.123441 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9615a643-2f07-4637-b8d0-3acb4f24d0d8-logs\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.123562 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-config-data\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.123588 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-config-data-custom\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.123616 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxhg5\" (UniqueName: \"kubernetes.io/projected/9615a643-2f07-4637-b8d0-3acb4f24d0d8-kube-api-access-jxhg5\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.123640 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-combined-ca-bundle\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.224838 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-public-tls-certs\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.225189 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9615a643-2f07-4637-b8d0-3acb4f24d0d8-logs\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.225350 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-config-data\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.225469 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-config-data-custom\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.225595 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxhg5\" (UniqueName: \"kubernetes.io/projected/9615a643-2f07-4637-b8d0-3acb4f24d0d8-kube-api-access-jxhg5\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.225711 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-combined-ca-bundle\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.225864 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-internal-tls-certs\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.228200 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9615a643-2f07-4637-b8d0-3acb4f24d0d8-logs\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.254003 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-config-data-custom\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.254803 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxhg5\" (UniqueName: \"kubernetes.io/projected/9615a643-2f07-4637-b8d0-3acb4f24d0d8-kube-api-access-jxhg5\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.263827 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-combined-ca-bundle\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.264806 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-public-tls-certs\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.264816 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-internal-tls-certs\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.265919 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9615a643-2f07-4637-b8d0-3acb4f24d0d8-config-data\") pod \"barbican-api-7f6f97bcf4-qbb2j\" (UID: \"9615a643-2f07-4637-b8d0-3acb4f24d0d8\") " pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.343463 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.819004 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f6f97bcf4-qbb2j"] Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.999415 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f6f97bcf4-qbb2j" event={"ID":"9615a643-2f07-4637-b8d0-3acb4f24d0d8","Type":"ContainerStarted","Data":"ad9d8ea1f52c786f2d5e6c61623060c11299829de1e123b057f7e2d2fc85fe7e"} Oct 14 15:33:08 crc kubenswrapper[4945]: I1014 15:33:08.999459 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:09 crc kubenswrapper[4945]: I1014 15:33:09.268859 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:33:09 crc kubenswrapper[4945]: I1014 15:33:09.418209 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:33:09 crc kubenswrapper[4945]: I1014 15:33:09.482600 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-8jh4x"] Oct 14 15:33:09 crc kubenswrapper[4945]: I1014 15:33:09.482819 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" podUID="b3e29775-9dfb-4b3b-b48a-b0464c5c296d" containerName="dnsmasq-dns" containerID="cri-o://bfef77538cfe55588ec4c4d03bbb860a9210f04403ba79e5aac684993b6a79ce" gracePeriod=10 Oct 14 15:33:09 crc kubenswrapper[4945]: I1014 15:33:09.847763 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" podUID="b3e29775-9dfb-4b3b-b48a-b0464c5c296d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: connect: connection refused" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.036251 4945 generic.go:334] "Generic (PLEG): container finished" podID="b3e29775-9dfb-4b3b-b48a-b0464c5c296d" containerID="bfef77538cfe55588ec4c4d03bbb860a9210f04403ba79e5aac684993b6a79ce" exitCode=0 Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.036386 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" event={"ID":"b3e29775-9dfb-4b3b-b48a-b0464c5c296d","Type":"ContainerDied","Data":"bfef77538cfe55588ec4c4d03bbb860a9210f04403ba79e5aac684993b6a79ce"} Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.054574 4945 generic.go:334] "Generic (PLEG): container finished" podID="367d71fe-5d2e-470a-958d-4a97ad15545c" containerID="dcc40ef21d557fe8592144e3e06558651617d428785e6cfc77ff5b21a8ffbf02" exitCode=0 Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.054752 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-gcqrl" event={"ID":"367d71fe-5d2e-470a-958d-4a97ad15545c","Type":"ContainerDied","Data":"dcc40ef21d557fe8592144e3e06558651617d428785e6cfc77ff5b21a8ffbf02"} Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.073632 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.081257 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f6f97bcf4-qbb2j" event={"ID":"9615a643-2f07-4637-b8d0-3acb4f24d0d8","Type":"ContainerStarted","Data":"f25edc7153797a0dea7ca780df033fe11d3f17416289ab5ecdf024b44e5f7e9e"} Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.081312 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f6f97bcf4-qbb2j" event={"ID":"9615a643-2f07-4637-b8d0-3acb4f24d0d8","Type":"ContainerStarted","Data":"8d297647ea1fdafe95b64903d5bff980b3bd624b3649aec7da75653746dc1e28"} Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.081909 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.082085 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.086428 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b210a0b1-468b-4e67-8618-02cb27c78033","Type":"ContainerStarted","Data":"1a805eaddb1a55fa2a7622701e5676ba6733c7a2e6db09fb7a45a6691b38ae24"} Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.086463 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.150493 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7f6f97bcf4-qbb2j" podStartSLOduration=3.1504733639999998 podStartE2EDuration="3.150473364s" podCreationTimestamp="2025-10-14 15:33:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:33:10.130787771 +0000 UTC m=+1120.114836139" watchObservedRunningTime="2025-10-14 15:33:10.150473364 +0000 UTC m=+1120.134521732" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.153529 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.202913 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.215764 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.215748762 podStartE2EDuration="5.215748762s" podCreationTimestamp="2025-10-14 15:33:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:33:10.209531024 +0000 UTC m=+1120.193579392" watchObservedRunningTime="2025-10-14 15:33:10.215748762 +0000 UTC m=+1120.199797120" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.345948 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.383763 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-ovsdbserver-sb\") pod \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.383841 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-ovsdbserver-nb\") pod \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.383891 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-dns-svc\") pod \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.384055 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-dns-swift-storage-0\") pod \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.384166 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwsb2\" (UniqueName: \"kubernetes.io/projected/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-kube-api-access-dwsb2\") pod \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.384242 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-config\") pod \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\" (UID: \"b3e29775-9dfb-4b3b-b48a-b0464c5c296d\") " Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.418757 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-kube-api-access-dwsb2" (OuterVolumeSpecName: "kube-api-access-dwsb2") pod "b3e29775-9dfb-4b3b-b48a-b0464c5c296d" (UID: "b3e29775-9dfb-4b3b-b48a-b0464c5c296d"). InnerVolumeSpecName "kube-api-access-dwsb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.475118 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b3e29775-9dfb-4b3b-b48a-b0464c5c296d" (UID: "b3e29775-9dfb-4b3b-b48a-b0464c5c296d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.476933 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-config" (OuterVolumeSpecName: "config") pod "b3e29775-9dfb-4b3b-b48a-b0464c5c296d" (UID: "b3e29775-9dfb-4b3b-b48a-b0464c5c296d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.477288 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b3e29775-9dfb-4b3b-b48a-b0464c5c296d" (UID: "b3e29775-9dfb-4b3b-b48a-b0464c5c296d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.494549 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.494594 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.494607 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwsb2\" (UniqueName: \"kubernetes.io/projected/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-kube-api-access-dwsb2\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.494617 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.520208 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b3e29775-9dfb-4b3b-b48a-b0464c5c296d" (UID: "b3e29775-9dfb-4b3b-b48a-b0464c5c296d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.535754 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b3e29775-9dfb-4b3b-b48a-b0464c5c296d" (UID: "b3e29775-9dfb-4b3b-b48a-b0464c5c296d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.596913 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:10 crc kubenswrapper[4945]: I1014 15:33:10.596959 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3e29775-9dfb-4b3b-b48a-b0464c5c296d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:11 crc kubenswrapper[4945]: I1014 15:33:11.115924 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" event={"ID":"b3e29775-9dfb-4b3b-b48a-b0464c5c296d","Type":"ContainerDied","Data":"ed808ca153382d113a6eaa2a3e32bb9ddad57e0eb96182c3ed8274efe5004bce"} Oct 14 15:33:11 crc kubenswrapper[4945]: I1014 15:33:11.117043 4945 scope.go:117] "RemoveContainer" containerID="bfef77538cfe55588ec4c4d03bbb860a9210f04403ba79e5aac684993b6a79ce" Oct 14 15:33:11 crc kubenswrapper[4945]: I1014 15:33:11.116297 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-8jh4x" Oct 14 15:33:11 crc kubenswrapper[4945]: I1014 15:33:11.126316 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-gcqrl" event={"ID":"367d71fe-5d2e-470a-958d-4a97ad15545c","Type":"ContainerStarted","Data":"0862d85cb4e434708c8ab5afe77dd2faba5f600c5bc5edc34dd1f5983075676b"} Oct 14 15:33:11 crc kubenswrapper[4945]: I1014 15:33:11.127299 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="82646027-6477-4a53-bbcc-cdf19a9e03da" containerName="cinder-scheduler" containerID="cri-o://f9269ae02ee5b157b988e20f7d2940b824ee467a8bc14eef9bf251081291b6be" gracePeriod=30 Oct 14 15:33:11 crc kubenswrapper[4945]: I1014 15:33:11.127966 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="82646027-6477-4a53-bbcc-cdf19a9e03da" containerName="probe" containerID="cri-o://1dea1f38fddd12ad1d406319300028dfceaa3449363ead51e289edeacc699c94" gracePeriod=30 Oct 14 15:33:11 crc kubenswrapper[4945]: I1014 15:33:11.152316 4945 scope.go:117] "RemoveContainer" containerID="a7171f5e54418c36468ddbdd19621bfff73e2b28cfb419d6591a5eddff1c22ba" Oct 14 15:33:11 crc kubenswrapper[4945]: I1014 15:33:11.180922 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-8jh4x"] Oct 14 15:33:11 crc kubenswrapper[4945]: I1014 15:33:11.201893 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-8jh4x"] Oct 14 15:33:11 crc kubenswrapper[4945]: I1014 15:33:11.201942 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-db-sync-gcqrl" podStartSLOduration=13.956500724 podStartE2EDuration="44.201919656s" podCreationTimestamp="2025-10-14 15:32:27 +0000 UTC" firstStartedPulling="2025-10-14 15:32:39.032848548 +0000 UTC m=+1089.016896936" lastFinishedPulling="2025-10-14 15:33:09.2782675 +0000 UTC m=+1119.262315868" observedRunningTime="2025-10-14 15:33:11.19156973 +0000 UTC m=+1121.175618098" watchObservedRunningTime="2025-10-14 15:33:11.201919656 +0000 UTC m=+1121.185968024" Oct 14 15:33:12 crc kubenswrapper[4945]: I1014 15:33:12.138832 4945 generic.go:334] "Generic (PLEG): container finished" podID="82646027-6477-4a53-bbcc-cdf19a9e03da" containerID="1dea1f38fddd12ad1d406319300028dfceaa3449363ead51e289edeacc699c94" exitCode=0 Oct 14 15:33:12 crc kubenswrapper[4945]: I1014 15:33:12.138946 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"82646027-6477-4a53-bbcc-cdf19a9e03da","Type":"ContainerDied","Data":"1dea1f38fddd12ad1d406319300028dfceaa3449363ead51e289edeacc699c94"} Oct 14 15:33:12 crc kubenswrapper[4945]: I1014 15:33:12.182813 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5cdf875b4d-nbblw" Oct 14 15:33:12 crc kubenswrapper[4945]: I1014 15:33:12.773146 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3e29775-9dfb-4b3b-b48a-b0464c5c296d" path="/var/lib/kubelet/pods/b3e29775-9dfb-4b3b-b48a-b0464c5c296d/volumes" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.158707 4945 generic.go:334] "Generic (PLEG): container finished" podID="82646027-6477-4a53-bbcc-cdf19a9e03da" containerID="f9269ae02ee5b157b988e20f7d2940b824ee467a8bc14eef9bf251081291b6be" exitCode=0 Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.158759 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"82646027-6477-4a53-bbcc-cdf19a9e03da","Type":"ContainerDied","Data":"f9269ae02ee5b157b988e20f7d2940b824ee467a8bc14eef9bf251081291b6be"} Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.391929 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.581147 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfsmr\" (UniqueName: \"kubernetes.io/projected/82646027-6477-4a53-bbcc-cdf19a9e03da-kube-api-access-bfsmr\") pod \"82646027-6477-4a53-bbcc-cdf19a9e03da\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.581359 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-scripts\") pod \"82646027-6477-4a53-bbcc-cdf19a9e03da\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.581388 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-combined-ca-bundle\") pod \"82646027-6477-4a53-bbcc-cdf19a9e03da\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.581440 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82646027-6477-4a53-bbcc-cdf19a9e03da-etc-machine-id\") pod \"82646027-6477-4a53-bbcc-cdf19a9e03da\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.581485 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-config-data-custom\") pod \"82646027-6477-4a53-bbcc-cdf19a9e03da\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.581549 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-config-data\") pod \"82646027-6477-4a53-bbcc-cdf19a9e03da\" (UID: \"82646027-6477-4a53-bbcc-cdf19a9e03da\") " Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.581552 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82646027-6477-4a53-bbcc-cdf19a9e03da-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "82646027-6477-4a53-bbcc-cdf19a9e03da" (UID: "82646027-6477-4a53-bbcc-cdf19a9e03da"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.582031 4945 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82646027-6477-4a53-bbcc-cdf19a9e03da-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.589055 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-scripts" (OuterVolumeSpecName: "scripts") pod "82646027-6477-4a53-bbcc-cdf19a9e03da" (UID: "82646027-6477-4a53-bbcc-cdf19a9e03da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.589119 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82646027-6477-4a53-bbcc-cdf19a9e03da-kube-api-access-bfsmr" (OuterVolumeSpecName: "kube-api-access-bfsmr") pod "82646027-6477-4a53-bbcc-cdf19a9e03da" (UID: "82646027-6477-4a53-bbcc-cdf19a9e03da"). InnerVolumeSpecName "kube-api-access-bfsmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.589184 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "82646027-6477-4a53-bbcc-cdf19a9e03da" (UID: "82646027-6477-4a53-bbcc-cdf19a9e03da"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.662979 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82646027-6477-4a53-bbcc-cdf19a9e03da" (UID: "82646027-6477-4a53-bbcc-cdf19a9e03da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.684328 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfsmr\" (UniqueName: \"kubernetes.io/projected/82646027-6477-4a53-bbcc-cdf19a9e03da-kube-api-access-bfsmr\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.684369 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.684380 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.684390 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.711602 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-config-data" (OuterVolumeSpecName: "config-data") pod "82646027-6477-4a53-bbcc-cdf19a9e03da" (UID: "82646027-6477-4a53-bbcc-cdf19a9e03da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:14 crc kubenswrapper[4945]: I1014 15:33:14.789041 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82646027-6477-4a53-bbcc-cdf19a9e03da-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.174404 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"82646027-6477-4a53-bbcc-cdf19a9e03da","Type":"ContainerDied","Data":"024b4748b9703b10aa1f05cd033ba09d61b3a10d6227661b49539bba76a6a9e4"} Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.174715 4945 scope.go:117] "RemoveContainer" containerID="1dea1f38fddd12ad1d406319300028dfceaa3449363ead51e289edeacc699c94" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.174857 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.207993 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.217785 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.234617 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 15:33:15 crc kubenswrapper[4945]: E1014 15:33:15.235003 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e29775-9dfb-4b3b-b48a-b0464c5c296d" containerName="init" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.235019 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e29775-9dfb-4b3b-b48a-b0464c5c296d" containerName="init" Oct 14 15:33:15 crc kubenswrapper[4945]: E1014 15:33:15.235038 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82646027-6477-4a53-bbcc-cdf19a9e03da" containerName="cinder-scheduler" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.235036 4945 scope.go:117] "RemoveContainer" containerID="f9269ae02ee5b157b988e20f7d2940b824ee467a8bc14eef9bf251081291b6be" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.235045 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="82646027-6477-4a53-bbcc-cdf19a9e03da" containerName="cinder-scheduler" Oct 14 15:33:15 crc kubenswrapper[4945]: E1014 15:33:15.235193 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82646027-6477-4a53-bbcc-cdf19a9e03da" containerName="probe" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.235206 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="82646027-6477-4a53-bbcc-cdf19a9e03da" containerName="probe" Oct 14 15:33:15 crc kubenswrapper[4945]: E1014 15:33:15.235227 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e29775-9dfb-4b3b-b48a-b0464c5c296d" containerName="dnsmasq-dns" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.235233 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e29775-9dfb-4b3b-b48a-b0464c5c296d" containerName="dnsmasq-dns" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.235503 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e29775-9dfb-4b3b-b48a-b0464c5c296d" containerName="dnsmasq-dns" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.235522 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="82646027-6477-4a53-bbcc-cdf19a9e03da" containerName="probe" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.235535 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="82646027-6477-4a53-bbcc-cdf19a9e03da" containerName="cinder-scheduler" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.236446 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.240331 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.257313 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.399101 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhr7f\" (UniqueName: \"kubernetes.io/projected/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-kube-api-access-vhr7f\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.399805 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.400148 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.400195 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-scripts\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.400240 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-config-data\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.400326 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.501467 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-config-data\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.501513 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.501560 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhr7f\" (UniqueName: \"kubernetes.io/projected/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-kube-api-access-vhr7f\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.501644 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.501731 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-scripts\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.501745 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.501800 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.505842 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-scripts\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.506754 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-config-data\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.507420 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.516645 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.518842 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhr7f\" (UniqueName: \"kubernetes.io/projected/b15e8808-64a4-4e83-8a8a-1ff4ecc68711-kube-api-access-vhr7f\") pod \"cinder-scheduler-0\" (UID: \"b15e8808-64a4-4e83-8a8a-1ff4ecc68711\") " pod="openstack/cinder-scheduler-0" Oct 14 15:33:15 crc kubenswrapper[4945]: I1014 15:33:15.565156 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.088755 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 15:33:16 crc kubenswrapper[4945]: W1014 15:33:16.095853 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb15e8808_64a4_4e83_8a8a_1ff4ecc68711.slice/crio-57f4220f0958f386c60cd6425e0b34dc3a93f392b9b0523f5cbb626f523b7be9 WatchSource:0}: Error finding container 57f4220f0958f386c60cd6425e0b34dc3a93f392b9b0523f5cbb626f523b7be9: Status 404 returned error can't find the container with id 57f4220f0958f386c60cd6425e0b34dc3a93f392b9b0523f5cbb626f523b7be9 Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.195938 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.197378 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.209650 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.211559 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.211791 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.211939 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-sxfps" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.215670 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95bm7\" (UniqueName: \"kubernetes.io/projected/473a83ad-e7a6-4b5b-899e-b64ef41465c9-kube-api-access-95bm7\") pod \"openstackclient\" (UID: \"473a83ad-e7a6-4b5b-899e-b64ef41465c9\") " pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.215725 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473a83ad-e7a6-4b5b-899e-b64ef41465c9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"473a83ad-e7a6-4b5b-899e-b64ef41465c9\") " pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.215751 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/473a83ad-e7a6-4b5b-899e-b64ef41465c9-openstack-config-secret\") pod \"openstackclient\" (UID: \"473a83ad-e7a6-4b5b-899e-b64ef41465c9\") " pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.215794 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/473a83ad-e7a6-4b5b-899e-b64ef41465c9-openstack-config\") pod \"openstackclient\" (UID: \"473a83ad-e7a6-4b5b-899e-b64ef41465c9\") " pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.265450 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b15e8808-64a4-4e83-8a8a-1ff4ecc68711","Type":"ContainerStarted","Data":"57f4220f0958f386c60cd6425e0b34dc3a93f392b9b0523f5cbb626f523b7be9"} Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.319854 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473a83ad-e7a6-4b5b-899e-b64ef41465c9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"473a83ad-e7a6-4b5b-899e-b64ef41465c9\") " pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.319930 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/473a83ad-e7a6-4b5b-899e-b64ef41465c9-openstack-config-secret\") pod \"openstackclient\" (UID: \"473a83ad-e7a6-4b5b-899e-b64ef41465c9\") " pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.319978 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/473a83ad-e7a6-4b5b-899e-b64ef41465c9-openstack-config\") pod \"openstackclient\" (UID: \"473a83ad-e7a6-4b5b-899e-b64ef41465c9\") " pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.320064 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95bm7\" (UniqueName: \"kubernetes.io/projected/473a83ad-e7a6-4b5b-899e-b64ef41465c9-kube-api-access-95bm7\") pod \"openstackclient\" (UID: \"473a83ad-e7a6-4b5b-899e-b64ef41465c9\") " pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.325259 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/473a83ad-e7a6-4b5b-899e-b64ef41465c9-openstack-config\") pod \"openstackclient\" (UID: \"473a83ad-e7a6-4b5b-899e-b64ef41465c9\") " pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.337571 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473a83ad-e7a6-4b5b-899e-b64ef41465c9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"473a83ad-e7a6-4b5b-899e-b64ef41465c9\") " pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.368401 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/473a83ad-e7a6-4b5b-899e-b64ef41465c9-openstack-config-secret\") pod \"openstackclient\" (UID: \"473a83ad-e7a6-4b5b-899e-b64ef41465c9\") " pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.393458 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95bm7\" (UniqueName: \"kubernetes.io/projected/473a83ad-e7a6-4b5b-899e-b64ef41465c9-kube-api-access-95bm7\") pod \"openstackclient\" (UID: \"473a83ad-e7a6-4b5b-899e-b64ef41465c9\") " pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.437628 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.783398 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82646027-6477-4a53-bbcc-cdf19a9e03da" path="/var/lib/kubelet/pods/82646027-6477-4a53-bbcc-cdf19a9e03da/volumes" Oct 14 15:33:16 crc kubenswrapper[4945]: I1014 15:33:16.940162 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 15:33:16 crc kubenswrapper[4945]: W1014 15:33:16.949001 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod473a83ad_e7a6_4b5b_899e_b64ef41465c9.slice/crio-f72899f15d7d108a7f1850dded53ea802e9cbef507db21507d18292be8b444f8 WatchSource:0}: Error finding container f72899f15d7d108a7f1850dded53ea802e9cbef507db21507d18292be8b444f8: Status 404 returned error can't find the container with id f72899f15d7d108a7f1850dded53ea802e9cbef507db21507d18292be8b444f8 Oct 14 15:33:17 crc kubenswrapper[4945]: I1014 15:33:17.305213 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"473a83ad-e7a6-4b5b-899e-b64ef41465c9","Type":"ContainerStarted","Data":"f72899f15d7d108a7f1850dded53ea802e9cbef507db21507d18292be8b444f8"} Oct 14 15:33:17 crc kubenswrapper[4945]: I1014 15:33:17.316773 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b15e8808-64a4-4e83-8a8a-1ff4ecc68711","Type":"ContainerStarted","Data":"b2bda44d689c922feba149ce2f03e7d7f2bc4c7dd9a6f275f17665af1d3f0265"} Oct 14 15:33:18 crc kubenswrapper[4945]: I1014 15:33:18.335514 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b15e8808-64a4-4e83-8a8a-1ff4ecc68711","Type":"ContainerStarted","Data":"98767712e5e8fb5e7a0c6af317af0927c30c2a72805fe61bfbd078a880b89b88"} Oct 14 15:33:18 crc kubenswrapper[4945]: I1014 15:33:18.366738 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.366714741 podStartE2EDuration="3.366714741s" podCreationTimestamp="2025-10-14 15:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:33:18.35477805 +0000 UTC m=+1128.338826418" watchObservedRunningTime="2025-10-14 15:33:18.366714741 +0000 UTC m=+1128.350763109" Oct 14 15:33:18 crc kubenswrapper[4945]: I1014 15:33:18.650482 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 14 15:33:20 crc kubenswrapper[4945]: I1014 15:33:20.138662 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:20 crc kubenswrapper[4945]: I1014 15:33:20.545846 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f6f97bcf4-qbb2j" Oct 14 15:33:20 crc kubenswrapper[4945]: I1014 15:33:20.566251 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 14 15:33:20 crc kubenswrapper[4945]: I1014 15:33:20.610737 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57988bf6fb-7btwh"] Oct 14 15:33:20 crc kubenswrapper[4945]: I1014 15:33:20.614023 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57988bf6fb-7btwh" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerName="barbican-api-log" containerID="cri-o://7ca2005fe6847cee8addd7898853ce5a01c638da8a163d7854d623dc05eda703" gracePeriod=30 Oct 14 15:33:20 crc kubenswrapper[4945]: I1014 15:33:20.614496 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57988bf6fb-7btwh" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerName="barbican-api" containerID="cri-o://b5f7ac0ecd4067d404414c5a558be63bed3ccc4acd14b1957f0c1cf1fae0b94b" gracePeriod=30 Oct 14 15:33:21 crc kubenswrapper[4945]: I1014 15:33:21.368130 4945 generic.go:334] "Generic (PLEG): container finished" podID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerID="7ca2005fe6847cee8addd7898853ce5a01c638da8a163d7854d623dc05eda703" exitCode=143 Oct 14 15:33:21 crc kubenswrapper[4945]: I1014 15:33:21.368466 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57988bf6fb-7btwh" event={"ID":"f93b99df-9b1f-42de-9da1-e537156ec1e9","Type":"ContainerDied","Data":"7ca2005fe6847cee8addd7898853ce5a01c638da8a163d7854d623dc05eda703"} Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.042681 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-86b568845c-dgcjh"] Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.045705 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.048145 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.048372 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.048526 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.059997 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-86b568845c-dgcjh"] Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.147585 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xjmz\" (UniqueName: \"kubernetes.io/projected/047520e2-5898-42bd-85bf-dce4063f08f4-kube-api-access-6xjmz\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.147645 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047520e2-5898-42bd-85bf-dce4063f08f4-config-data\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.147890 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/047520e2-5898-42bd-85bf-dce4063f08f4-public-tls-certs\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.147938 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/047520e2-5898-42bd-85bf-dce4063f08f4-etc-swift\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.147997 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/047520e2-5898-42bd-85bf-dce4063f08f4-internal-tls-certs\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.148072 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047520e2-5898-42bd-85bf-dce4063f08f4-log-httpd\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.148098 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047520e2-5898-42bd-85bf-dce4063f08f4-run-httpd\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.148126 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047520e2-5898-42bd-85bf-dce4063f08f4-combined-ca-bundle\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.155867 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.156126 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="ceilometer-central-agent" containerID="cri-o://ddbba1cf3884c0569f9d8f731c0f13f71e0f040a7b2e85d865178c53f8d2b484" gracePeriod=30 Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.156795 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="sg-core" containerID="cri-o://e466723f709f1ff3739636b9b6ee0f4ccaf3c5a203968a48cf2c38df46d2777b" gracePeriod=30 Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.156811 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="proxy-httpd" containerID="cri-o://8a3426b80153b65b4a2094de34d3d5b9ae3026558616871dfe038a68769a0ca7" gracePeriod=30 Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.156862 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="ceilometer-notification-agent" containerID="cri-o://aa819694cf6724c3d6eda56f7d1dcda666b4d83a95af317fef6782410af75377" gracePeriod=30 Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.166682 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.251811 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/047520e2-5898-42bd-85bf-dce4063f08f4-public-tls-certs\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.251863 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/047520e2-5898-42bd-85bf-dce4063f08f4-etc-swift\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.251918 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/047520e2-5898-42bd-85bf-dce4063f08f4-internal-tls-certs\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.251969 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047520e2-5898-42bd-85bf-dce4063f08f4-log-httpd\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.251996 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047520e2-5898-42bd-85bf-dce4063f08f4-run-httpd\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.252036 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047520e2-5898-42bd-85bf-dce4063f08f4-combined-ca-bundle\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.252064 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xjmz\" (UniqueName: \"kubernetes.io/projected/047520e2-5898-42bd-85bf-dce4063f08f4-kube-api-access-6xjmz\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.252103 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047520e2-5898-42bd-85bf-dce4063f08f4-config-data\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.255358 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047520e2-5898-42bd-85bf-dce4063f08f4-log-httpd\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.257161 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047520e2-5898-42bd-85bf-dce4063f08f4-run-httpd\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.263827 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/047520e2-5898-42bd-85bf-dce4063f08f4-etc-swift\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.265754 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047520e2-5898-42bd-85bf-dce4063f08f4-config-data\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.267123 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/047520e2-5898-42bd-85bf-dce4063f08f4-public-tls-certs\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.269657 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/047520e2-5898-42bd-85bf-dce4063f08f4-internal-tls-certs\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.271332 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047520e2-5898-42bd-85bf-dce4063f08f4-combined-ca-bundle\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.272550 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xjmz\" (UniqueName: \"kubernetes.io/projected/047520e2-5898-42bd-85bf-dce4063f08f4-kube-api-access-6xjmz\") pod \"swift-proxy-86b568845c-dgcjh\" (UID: \"047520e2-5898-42bd-85bf-dce4063f08f4\") " pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.367502 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.378843 4945 generic.go:334] "Generic (PLEG): container finished" podID="367d71fe-5d2e-470a-958d-4a97ad15545c" containerID="0862d85cb4e434708c8ab5afe77dd2faba5f600c5bc5edc34dd1f5983075676b" exitCode=0 Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.378927 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-gcqrl" event={"ID":"367d71fe-5d2e-470a-958d-4a97ad15545c","Type":"ContainerDied","Data":"0862d85cb4e434708c8ab5afe77dd2faba5f600c5bc5edc34dd1f5983075676b"} Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.383555 4945 generic.go:334] "Generic (PLEG): container finished" podID="706a95cb-0318-4cad-af7a-04324bb19b23" containerID="e466723f709f1ff3739636b9b6ee0f4ccaf3c5a203968a48cf2c38df46d2777b" exitCode=2 Oct 14 15:33:22 crc kubenswrapper[4945]: I1014 15:33:22.383600 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"706a95cb-0318-4cad-af7a-04324bb19b23","Type":"ContainerDied","Data":"e466723f709f1ff3739636b9b6ee0f4ccaf3c5a203968a48cf2c38df46d2777b"} Oct 14 15:33:23 crc kubenswrapper[4945]: I1014 15:33:23.417510 4945 generic.go:334] "Generic (PLEG): container finished" podID="706a95cb-0318-4cad-af7a-04324bb19b23" containerID="8a3426b80153b65b4a2094de34d3d5b9ae3026558616871dfe038a68769a0ca7" exitCode=0 Oct 14 15:33:23 crc kubenswrapper[4945]: I1014 15:33:23.417789 4945 generic.go:334] "Generic (PLEG): container finished" podID="706a95cb-0318-4cad-af7a-04324bb19b23" containerID="aa819694cf6724c3d6eda56f7d1dcda666b4d83a95af317fef6782410af75377" exitCode=0 Oct 14 15:33:23 crc kubenswrapper[4945]: I1014 15:33:23.417797 4945 generic.go:334] "Generic (PLEG): container finished" podID="706a95cb-0318-4cad-af7a-04324bb19b23" containerID="ddbba1cf3884c0569f9d8f731c0f13f71e0f040a7b2e85d865178c53f8d2b484" exitCode=0 Oct 14 15:33:23 crc kubenswrapper[4945]: I1014 15:33:23.417974 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"706a95cb-0318-4cad-af7a-04324bb19b23","Type":"ContainerDied","Data":"8a3426b80153b65b4a2094de34d3d5b9ae3026558616871dfe038a68769a0ca7"} Oct 14 15:33:23 crc kubenswrapper[4945]: I1014 15:33:23.418033 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"706a95cb-0318-4cad-af7a-04324bb19b23","Type":"ContainerDied","Data":"aa819694cf6724c3d6eda56f7d1dcda666b4d83a95af317fef6782410af75377"} Oct 14 15:33:23 crc kubenswrapper[4945]: I1014 15:33:23.418045 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"706a95cb-0318-4cad-af7a-04324bb19b23","Type":"ContainerDied","Data":"ddbba1cf3884c0569f9d8f731c0f13f71e0f040a7b2e85d865178c53f8d2b484"} Oct 14 15:33:23 crc kubenswrapper[4945]: I1014 15:33:23.801249 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57988bf6fb-7btwh" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:33782->10.217.0.154:9311: read: connection reset by peer" Oct 14 15:33:23 crc kubenswrapper[4945]: I1014 15:33:23.801280 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57988bf6fb-7btwh" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:33766->10.217.0.154:9311: read: connection reset by peer" Oct 14 15:33:24 crc kubenswrapper[4945]: I1014 15:33:24.432641 4945 generic.go:334] "Generic (PLEG): container finished" podID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerID="b5f7ac0ecd4067d404414c5a558be63bed3ccc4acd14b1957f0c1cf1fae0b94b" exitCode=0 Oct 14 15:33:24 crc kubenswrapper[4945]: I1014 15:33:24.432738 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57988bf6fb-7btwh" event={"ID":"f93b99df-9b1f-42de-9da1-e537156ec1e9","Type":"ContainerDied","Data":"b5f7ac0ecd4067d404414c5a558be63bed3ccc4acd14b1957f0c1cf1fae0b94b"} Oct 14 15:33:25 crc kubenswrapper[4945]: I1014 15:33:25.820050 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 14 15:33:27 crc kubenswrapper[4945]: I1014 15:33:27.332734 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57988bf6fb-7btwh" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": dial tcp 10.217.0.154:9311: connect: connection refused" Oct 14 15:33:27 crc kubenswrapper[4945]: I1014 15:33:27.332734 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57988bf6fb-7btwh" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": dial tcp 10.217.0.154:9311: connect: connection refused" Oct 14 15:33:28 crc kubenswrapper[4945]: I1014 15:33:28.522060 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:33:28 crc kubenswrapper[4945]: I1014 15:33:28.525720 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-bd8df944d-q4z6c" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.287502 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.370276 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.490672 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.492711 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-gcqrl" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.492718 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-gcqrl" event={"ID":"367d71fe-5d2e-470a-958d-4a97ad15545c","Type":"ContainerDied","Data":"38e3c0946f017efb5322249021be0e4ab466cde2adcf1d7bc924dda8204aee85"} Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.492771 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38e3c0946f017efb5322249021be0e4ab466cde2adcf1d7bc924dda8204aee85" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.503747 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57988bf6fb-7btwh" event={"ID":"f93b99df-9b1f-42de-9da1-e537156ec1e9","Type":"ContainerDied","Data":"ed160bbb9270ff5eb638c65070f0601723e4ca267f722243004098693a553a7c"} Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.503806 4945 scope.go:117] "RemoveContainer" containerID="b5f7ac0ecd4067d404414c5a558be63bed3ccc4acd14b1957f0c1cf1fae0b94b" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.503982 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57988bf6fb-7btwh" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.520035 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"473a83ad-e7a6-4b5b-899e-b64ef41465c9","Type":"ContainerStarted","Data":"4d0ee32c54e6a9ea12b996209611c31df2c906940d01137bde36756758017023"} Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.535401 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-config-data\") pod \"367d71fe-5d2e-470a-958d-4a97ad15545c\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.535503 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/367d71fe-5d2e-470a-958d-4a97ad15545c-config-data-merged\") pod \"367d71fe-5d2e-470a-958d-4a97ad15545c\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.535574 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/367d71fe-5d2e-470a-958d-4a97ad15545c-etc-podinfo\") pod \"367d71fe-5d2e-470a-958d-4a97ad15545c\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.535641 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-scripts\") pod \"367d71fe-5d2e-470a-958d-4a97ad15545c\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.535665 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-combined-ca-bundle\") pod \"367d71fe-5d2e-470a-958d-4a97ad15545c\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.535732 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nt8br\" (UniqueName: \"kubernetes.io/projected/367d71fe-5d2e-470a-958d-4a97ad15545c-kube-api-access-nt8br\") pod \"367d71fe-5d2e-470a-958d-4a97ad15545c\" (UID: \"367d71fe-5d2e-470a-958d-4a97ad15545c\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.540790 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/367d71fe-5d2e-470a-958d-4a97ad15545c-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "367d71fe-5d2e-470a-958d-4a97ad15545c" (UID: "367d71fe-5d2e-470a-958d-4a97ad15545c"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.541099 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-scripts" (OuterVolumeSpecName: "scripts") pod "367d71fe-5d2e-470a-958d-4a97ad15545c" (UID: "367d71fe-5d2e-470a-958d-4a97ad15545c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.541503 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/367d71fe-5d2e-470a-958d-4a97ad15545c-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "367d71fe-5d2e-470a-958d-4a97ad15545c" (UID: "367d71fe-5d2e-470a-958d-4a97ad15545c"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.543982 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/367d71fe-5d2e-470a-958d-4a97ad15545c-kube-api-access-nt8br" (OuterVolumeSpecName: "kube-api-access-nt8br") pod "367d71fe-5d2e-470a-958d-4a97ad15545c" (UID: "367d71fe-5d2e-470a-958d-4a97ad15545c"). InnerVolumeSpecName "kube-api-access-nt8br". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.547148 4945 scope.go:117] "RemoveContainer" containerID="7ca2005fe6847cee8addd7898853ce5a01c638da8a163d7854d623dc05eda703" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.568134 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.32506539 podStartE2EDuration="13.568116674s" podCreationTimestamp="2025-10-14 15:33:16 +0000 UTC" firstStartedPulling="2025-10-14 15:33:16.951415379 +0000 UTC m=+1126.935463747" lastFinishedPulling="2025-10-14 15:33:29.194466663 +0000 UTC m=+1139.178515031" observedRunningTime="2025-10-14 15:33:29.533956786 +0000 UTC m=+1139.518005154" watchObservedRunningTime="2025-10-14 15:33:29.568116674 +0000 UTC m=+1139.552165032" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.571068 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-config-data" (OuterVolumeSpecName: "config-data") pod "367d71fe-5d2e-470a-958d-4a97ad15545c" (UID: "367d71fe-5d2e-470a-958d-4a97ad15545c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.610594 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "367d71fe-5d2e-470a-958d-4a97ad15545c" (UID: "367d71fe-5d2e-470a-958d-4a97ad15545c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.614670 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.642689 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-config-data\") pod \"f93b99df-9b1f-42de-9da1-e537156ec1e9\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.642765 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/706a95cb-0318-4cad-af7a-04324bb19b23-run-httpd\") pod \"706a95cb-0318-4cad-af7a-04324bb19b23\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.642789 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfm8n\" (UniqueName: \"kubernetes.io/projected/f93b99df-9b1f-42de-9da1-e537156ec1e9-kube-api-access-pfm8n\") pod \"f93b99df-9b1f-42de-9da1-e537156ec1e9\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.642827 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-config-data\") pod \"706a95cb-0318-4cad-af7a-04324bb19b23\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.642846 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsm4h\" (UniqueName: \"kubernetes.io/projected/706a95cb-0318-4cad-af7a-04324bb19b23-kube-api-access-dsm4h\") pod \"706a95cb-0318-4cad-af7a-04324bb19b23\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.642922 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-config-data-custom\") pod \"f93b99df-9b1f-42de-9da1-e537156ec1e9\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.642947 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/706a95cb-0318-4cad-af7a-04324bb19b23-log-httpd\") pod \"706a95cb-0318-4cad-af7a-04324bb19b23\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.643009 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-scripts\") pod \"706a95cb-0318-4cad-af7a-04324bb19b23\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.643035 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93b99df-9b1f-42de-9da1-e537156ec1e9-logs\") pod \"f93b99df-9b1f-42de-9da1-e537156ec1e9\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.643099 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-combined-ca-bundle\") pod \"706a95cb-0318-4cad-af7a-04324bb19b23\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.643126 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-combined-ca-bundle\") pod \"f93b99df-9b1f-42de-9da1-e537156ec1e9\" (UID: \"f93b99df-9b1f-42de-9da1-e537156ec1e9\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.643178 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-sg-core-conf-yaml\") pod \"706a95cb-0318-4cad-af7a-04324bb19b23\" (UID: \"706a95cb-0318-4cad-af7a-04324bb19b23\") " Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.643642 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.643665 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.643675 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nt8br\" (UniqueName: \"kubernetes.io/projected/367d71fe-5d2e-470a-958d-4a97ad15545c-kube-api-access-nt8br\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.643683 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367d71fe-5d2e-470a-958d-4a97ad15545c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.643710 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/367d71fe-5d2e-470a-958d-4a97ad15545c-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.643729 4945 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/367d71fe-5d2e-470a-958d-4a97ad15545c-etc-podinfo\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.654078 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f93b99df-9b1f-42de-9da1-e537156ec1e9-logs" (OuterVolumeSpecName: "logs") pod "f93b99df-9b1f-42de-9da1-e537156ec1e9" (UID: "f93b99df-9b1f-42de-9da1-e537156ec1e9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.654203 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/706a95cb-0318-4cad-af7a-04324bb19b23-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "706a95cb-0318-4cad-af7a-04324bb19b23" (UID: "706a95cb-0318-4cad-af7a-04324bb19b23"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.654539 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/706a95cb-0318-4cad-af7a-04324bb19b23-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "706a95cb-0318-4cad-af7a-04324bb19b23" (UID: "706a95cb-0318-4cad-af7a-04324bb19b23"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.658195 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f93b99df-9b1f-42de-9da1-e537156ec1e9-kube-api-access-pfm8n" (OuterVolumeSpecName: "kube-api-access-pfm8n") pod "f93b99df-9b1f-42de-9da1-e537156ec1e9" (UID: "f93b99df-9b1f-42de-9da1-e537156ec1e9"). InnerVolumeSpecName "kube-api-access-pfm8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.662404 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f93b99df-9b1f-42de-9da1-e537156ec1e9" (UID: "f93b99df-9b1f-42de-9da1-e537156ec1e9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.663551 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/706a95cb-0318-4cad-af7a-04324bb19b23-kube-api-access-dsm4h" (OuterVolumeSpecName: "kube-api-access-dsm4h") pod "706a95cb-0318-4cad-af7a-04324bb19b23" (UID: "706a95cb-0318-4cad-af7a-04324bb19b23"). InnerVolumeSpecName "kube-api-access-dsm4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.666201 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-scripts" (OuterVolumeSpecName: "scripts") pod "706a95cb-0318-4cad-af7a-04324bb19b23" (UID: "706a95cb-0318-4cad-af7a-04324bb19b23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.696540 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f93b99df-9b1f-42de-9da1-e537156ec1e9" (UID: "f93b99df-9b1f-42de-9da1-e537156ec1e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.706834 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.707097 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bb5c5b1d-4161-435d-8ebe-61016248b180" containerName="glance-log" containerID="cri-o://166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547" gracePeriod=30 Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.707897 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bb5c5b1d-4161-435d-8ebe-61016248b180" containerName="glance-httpd" containerID="cri-o://2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344" gracePeriod=30 Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.709727 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "706a95cb-0318-4cad-af7a-04324bb19b23" (UID: "706a95cb-0318-4cad-af7a-04324bb19b23"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.745234 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.750509 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/706a95cb-0318-4cad-af7a-04324bb19b23-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.750599 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.750701 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93b99df-9b1f-42de-9da1-e537156ec1e9-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.750775 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.750859 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.751056 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/706a95cb-0318-4cad-af7a-04324bb19b23-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.751124 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfm8n\" (UniqueName: \"kubernetes.io/projected/f93b99df-9b1f-42de-9da1-e537156ec1e9-kube-api-access-pfm8n\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.751191 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsm4h\" (UniqueName: \"kubernetes.io/projected/706a95cb-0318-4cad-af7a-04324bb19b23-kube-api-access-dsm4h\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.802054 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-config-data" (OuterVolumeSpecName: "config-data") pod "f93b99df-9b1f-42de-9da1-e537156ec1e9" (UID: "f93b99df-9b1f-42de-9da1-e537156ec1e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.833452 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "706a95cb-0318-4cad-af7a-04324bb19b23" (UID: "706a95cb-0318-4cad-af7a-04324bb19b23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.859286 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.859560 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93b99df-9b1f-42de-9da1-e537156ec1e9-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.927172 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-config-data" (OuterVolumeSpecName: "config-data") pod "706a95cb-0318-4cad-af7a-04324bb19b23" (UID: "706a95cb-0318-4cad-af7a-04324bb19b23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.963563 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/706a95cb-0318-4cad-af7a-04324bb19b23-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.978241 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57988bf6fb-7btwh"] Oct 14 15:33:29 crc kubenswrapper[4945]: I1014 15:33:29.992414 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-57988bf6fb-7btwh"] Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.025785 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-86b568845c-dgcjh"] Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.536494 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-86b568845c-dgcjh" event={"ID":"047520e2-5898-42bd-85bf-dce4063f08f4","Type":"ContainerStarted","Data":"6a99d297196cc7864b37bdfdde44f14da49e098fe4d2bb9fa0294b9216540dae"} Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.536737 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-86b568845c-dgcjh" event={"ID":"047520e2-5898-42bd-85bf-dce4063f08f4","Type":"ContainerStarted","Data":"37c66807fae9570cad648f7b0d25d926793443fd01fad8f8f5b765a60f0d9ab1"} Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.539346 4945 generic.go:334] "Generic (PLEG): container finished" podID="bb5c5b1d-4161-435d-8ebe-61016248b180" containerID="166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547" exitCode=143 Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.539417 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bb5c5b1d-4161-435d-8ebe-61016248b180","Type":"ContainerDied","Data":"166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547"} Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.543353 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"706a95cb-0318-4cad-af7a-04324bb19b23","Type":"ContainerDied","Data":"300fdb03106ab359adb493f73ea9942c25a4297c01ea32eb9d39f657633ba37e"} Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.543396 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.543435 4945 scope.go:117] "RemoveContainer" containerID="8a3426b80153b65b4a2094de34d3d5b9ae3026558616871dfe038a68769a0ca7" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.567560 4945 scope.go:117] "RemoveContainer" containerID="e466723f709f1ff3739636b9b6ee0f4ccaf3c5a203968a48cf2c38df46d2777b" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.588729 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.605809 4945 scope.go:117] "RemoveContainer" containerID="aa819694cf6724c3d6eda56f7d1dcda666b4d83a95af317fef6782410af75377" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.607121 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614098 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:30 crc kubenswrapper[4945]: E1014 15:33:30.614445 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="367d71fe-5d2e-470a-958d-4a97ad15545c" containerName="ironic-db-sync" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614461 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="367d71fe-5d2e-470a-958d-4a97ad15545c" containerName="ironic-db-sync" Oct 14 15:33:30 crc kubenswrapper[4945]: E1014 15:33:30.614473 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerName="barbican-api-log" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614481 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerName="barbican-api-log" Oct 14 15:33:30 crc kubenswrapper[4945]: E1014 15:33:30.614494 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerName="barbican-api" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614500 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerName="barbican-api" Oct 14 15:33:30 crc kubenswrapper[4945]: E1014 15:33:30.614516 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="ceilometer-notification-agent" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614523 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="ceilometer-notification-agent" Oct 14 15:33:30 crc kubenswrapper[4945]: E1014 15:33:30.614536 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="367d71fe-5d2e-470a-958d-4a97ad15545c" containerName="init" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614541 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="367d71fe-5d2e-470a-958d-4a97ad15545c" containerName="init" Oct 14 15:33:30 crc kubenswrapper[4945]: E1014 15:33:30.614555 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="sg-core" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614560 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="sg-core" Oct 14 15:33:30 crc kubenswrapper[4945]: E1014 15:33:30.614576 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="proxy-httpd" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614582 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="proxy-httpd" Oct 14 15:33:30 crc kubenswrapper[4945]: E1014 15:33:30.614594 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="ceilometer-central-agent" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614601 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="ceilometer-central-agent" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614772 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="367d71fe-5d2e-470a-958d-4a97ad15545c" containerName="ironic-db-sync" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614786 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerName="barbican-api" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614798 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="ceilometer-central-agent" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614809 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="ceilometer-notification-agent" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614817 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" containerName="barbican-api-log" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614829 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="sg-core" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.614838 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" containerName="proxy-httpd" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.621017 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.624692 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.625615 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.637572 4945 scope.go:117] "RemoveContainer" containerID="ddbba1cf3884c0569f9d8f731c0f13f71e0f040a7b2e85d865178c53f8d2b484" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.651857 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.679058 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-scripts\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.679302 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kckfx\" (UniqueName: \"kubernetes.io/projected/b18d41ba-3cf6-454f-ae01-bc9f35df1210-kube-api-access-kckfx\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.679378 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.679455 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b18d41ba-3cf6-454f-ae01-bc9f35df1210-run-httpd\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.679608 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-config-data\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.679740 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.679834 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b18d41ba-3cf6-454f-ae01-bc9f35df1210-log-httpd\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.780778 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-scripts\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.797425 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kckfx\" (UniqueName: \"kubernetes.io/projected/b18d41ba-3cf6-454f-ae01-bc9f35df1210-kube-api-access-kckfx\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.801105 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.801584 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b18d41ba-3cf6-454f-ae01-bc9f35df1210-run-httpd\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.806626 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-config-data\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.820611 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.820667 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.820705 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b18d41ba-3cf6-454f-ae01-bc9f35df1210-log-httpd\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.820507 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-config-data\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.804468 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-scripts\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.790805 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.806303 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.804667 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b18d41ba-3cf6-454f-ae01-bc9f35df1210-run-httpd\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.821364 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b18d41ba-3cf6-454f-ae01-bc9f35df1210-log-httpd\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.840582 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.841932 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kckfx\" (UniqueName: \"kubernetes.io/projected/b18d41ba-3cf6-454f-ae01-bc9f35df1210-kube-api-access-kckfx\") pod \"ceilometer-0\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " pod="openstack/ceilometer-0" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.853464 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="706a95cb-0318-4cad-af7a-04324bb19b23" path="/var/lib/kubelet/pods/706a95cb-0318-4cad-af7a-04324bb19b23/volumes" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.854366 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f93b99df-9b1f-42de-9da1-e537156ec1e9" path="/var/lib/kubelet/pods/f93b99df-9b1f-42de-9da1-e537156ec1e9/volumes" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.864609 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-db-create-8hrxc"] Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.866029 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-8hrxc" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.892252 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-create-8hrxc"] Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.897293 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-neutron-agent-648bd876cc-srkfr"] Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.898455 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.901377 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-dockercfg-g7lhl" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.901577 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-neutron-agent-config-data" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.922613 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv5g8\" (UniqueName: \"kubernetes.io/projected/92ed91ce-91c5-46ff-8e6c-959b9fa96552-kube-api-access-gv5g8\") pod \"ironic-inspector-db-create-8hrxc\" (UID: \"92ed91ce-91c5-46ff-8e6c-959b9fa96552\") " pod="openstack/ironic-inspector-db-create-8hrxc" Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.940648 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-neutron-agent-648bd876cc-srkfr"] Oct 14 15:33:30 crc kubenswrapper[4945]: I1014 15:33:30.945295 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.014892 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-78479b8b7d-pdhlw"] Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.020300 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.025211 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-api-scripts" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.025305 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-api-config-data" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.025542 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-config-data" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.025928 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/11506640-cea7-421c-878c-c4b449df880f-config\") pod \"ironic-neutron-agent-648bd876cc-srkfr\" (UID: \"11506640-cea7-421c-878c-c4b449df880f\") " pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.026090 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbs4j\" (UniqueName: \"kubernetes.io/projected/11506640-cea7-421c-878c-c4b449df880f-kube-api-access-rbs4j\") pod \"ironic-neutron-agent-648bd876cc-srkfr\" (UID: \"11506640-cea7-421c-878c-c4b449df880f\") " pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.026336 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv5g8\" (UniqueName: \"kubernetes.io/projected/92ed91ce-91c5-46ff-8e6c-959b9fa96552-kube-api-access-gv5g8\") pod \"ironic-inspector-db-create-8hrxc\" (UID: \"92ed91ce-91c5-46ff-8e6c-959b9fa96552\") " pod="openstack/ironic-inspector-db-create-8hrxc" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.026756 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11506640-cea7-421c-878c-c4b449df880f-combined-ca-bundle\") pod \"ironic-neutron-agent-648bd876cc-srkfr\" (UID: \"11506640-cea7-421c-878c-c4b449df880f\") " pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.060970 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-78479b8b7d-pdhlw"] Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.075536 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv5g8\" (UniqueName: \"kubernetes.io/projected/92ed91ce-91c5-46ff-8e6c-959b9fa96552-kube-api-access-gv5g8\") pod \"ironic-inspector-db-create-8hrxc\" (UID: \"92ed91ce-91c5-46ff-8e6c-959b9fa96552\") " pod="openstack/ironic-inspector-db-create-8hrxc" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.128690 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-logs\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.128750 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-etc-podinfo\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.128816 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data-merged\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.128859 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/11506640-cea7-421c-878c-c4b449df880f-config\") pod \"ironic-neutron-agent-648bd876cc-srkfr\" (UID: \"11506640-cea7-421c-878c-c4b449df880f\") " pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.128925 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-scripts\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.128965 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.128995 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbs4j\" (UniqueName: \"kubernetes.io/projected/11506640-cea7-421c-878c-c4b449df880f-kube-api-access-rbs4j\") pod \"ironic-neutron-agent-648bd876cc-srkfr\" (UID: \"11506640-cea7-421c-878c-c4b449df880f\") " pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.129055 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-combined-ca-bundle\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.129098 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11506640-cea7-421c-878c-c4b449df880f-combined-ca-bundle\") pod \"ironic-neutron-agent-648bd876cc-srkfr\" (UID: \"11506640-cea7-421c-878c-c4b449df880f\") " pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.129138 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data-custom\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.129179 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbgtp\" (UniqueName: \"kubernetes.io/projected/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-kube-api-access-sbgtp\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.136077 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11506640-cea7-421c-878c-c4b449df880f-combined-ca-bundle\") pod \"ironic-neutron-agent-648bd876cc-srkfr\" (UID: \"11506640-cea7-421c-878c-c4b449df880f\") " pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.136560 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/11506640-cea7-421c-878c-c4b449df880f-config\") pod \"ironic-neutron-agent-648bd876cc-srkfr\" (UID: \"11506640-cea7-421c-878c-c4b449df880f\") " pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.154056 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbs4j\" (UniqueName: \"kubernetes.io/projected/11506640-cea7-421c-878c-c4b449df880f-kube-api-access-rbs4j\") pod \"ironic-neutron-agent-648bd876cc-srkfr\" (UID: \"11506640-cea7-421c-878c-c4b449df880f\") " pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.214411 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-8hrxc" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.231970 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-logs\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.232013 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-etc-podinfo\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.232070 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data-merged\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.232134 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-scripts\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.232176 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.232243 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-combined-ca-bundle\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.232389 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data-custom\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.232430 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbgtp\" (UniqueName: \"kubernetes.io/projected/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-kube-api-access-sbgtp\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.233631 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data-merged\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.232390 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-logs\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.238493 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-scripts\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.239225 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-combined-ca-bundle\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.240601 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.244395 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-etc-podinfo\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.246464 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data-custom\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.256039 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbgtp\" (UniqueName: \"kubernetes.io/projected/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-kube-api-access-sbgtp\") pod \"ironic-78479b8b7d-pdhlw\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.266119 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.440388 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.547712 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.595664 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-86b568845c-dgcjh" event={"ID":"047520e2-5898-42bd-85bf-dce4063f08f4","Type":"ContainerStarted","Data":"6eb742ebca96621f087eee3e003d3ba736d2742dd5f8f42788769a3f3acbbb5a"} Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.596154 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.596175 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.624404 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-86b568845c-dgcjh" podStartSLOduration=9.624377573 podStartE2EDuration="9.624377573s" podCreationTimestamp="2025-10-14 15:33:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:33:31.624186787 +0000 UTC m=+1141.608235155" watchObservedRunningTime="2025-10-14 15:33:31.624377573 +0000 UTC m=+1141.608425941" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.794734 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-create-8hrxc"] Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.844970 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-conductor-0"] Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.847647 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-conductor-0" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.852344 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-conductor-scripts" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.852392 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-conductor-config-data" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.869503 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-conductor-0"] Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.925318 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-neutron-agent-648bd876cc-srkfr"] Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.954939 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.955139 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-config-data\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.955248 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.955391 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x2rh\" (UniqueName: \"kubernetes.io/projected/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-kube-api-access-8x2rh\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.955435 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.955610 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.955632 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-scripts\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:31 crc kubenswrapper[4945]: I1014 15:33:31.955703 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.023576 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-78479b8b7d-pdhlw"] Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.057030 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.057128 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.057176 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-config-data\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.057220 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.057293 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x2rh\" (UniqueName: \"kubernetes.io/projected/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-kube-api-access-8x2rh\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.057324 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.057429 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.057452 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-scripts\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.058399 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.059354 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.068672 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.071979 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.097275 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-scripts\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.098764 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x2rh\" (UniqueName: \"kubernetes.io/projected/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-kube-api-access-8x2rh\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.118364 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-config-data\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.125851 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/042f4fc3-10ec-450e-b3cf-77a11d71a8b7-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.206523 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ironic-conductor-0\" (UID: \"042f4fc3-10ec-450e-b3cf-77a11d71a8b7\") " pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.232136 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.479382 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-conductor-0" Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.663445 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b18d41ba-3cf6-454f-ae01-bc9f35df1210","Type":"ContainerStarted","Data":"2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9"} Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.663497 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b18d41ba-3cf6-454f-ae01-bc9f35df1210","Type":"ContainerStarted","Data":"b332e70687114937a0ab9a38b755ee566e1b736b8cdaf87324e18fd3ddc794eb"} Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.672255 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-78479b8b7d-pdhlw" event={"ID":"5a183ba7-f9df-45a6-8f41-43e4d16eebdd","Type":"ContainerStarted","Data":"afa08e7038da40a958ca7bc580c1ab7f1f231dad3ea9bc61e6979463e772698a"} Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.678027 4945 generic.go:334] "Generic (PLEG): container finished" podID="92ed91ce-91c5-46ff-8e6c-959b9fa96552" containerID="5f0cf2932656c6e5688adbf9fac5a3175e98d84306a5935b0046850b86a9e240" exitCode=0 Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.678105 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-8hrxc" event={"ID":"92ed91ce-91c5-46ff-8e6c-959b9fa96552","Type":"ContainerDied","Data":"5f0cf2932656c6e5688adbf9fac5a3175e98d84306a5935b0046850b86a9e240"} Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.678131 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-8hrxc" event={"ID":"92ed91ce-91c5-46ff-8e6c-959b9fa96552","Type":"ContainerStarted","Data":"75bb16074604883801ff2dbf80d9e4fa9f2ba408e1208f872acec9d0f682963b"} Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.686994 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" event={"ID":"11506640-cea7-421c-878c-c4b449df880f","Type":"ContainerStarted","Data":"25a774b1375cfde57f231bc14d775dc04cea77912665f90bc8a1260ee50da3d7"} Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.911792 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.915792 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="aa12f709-e635-4685-a892-4c65ba588f36" containerName="glance-httpd" containerID="cri-o://b241e8b17c3ea5dd6f7019383c566e00ac84c1af7a799b95ea7ad241b3f1a0a8" gracePeriod=30 Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.916208 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="aa12f709-e635-4685-a892-4c65ba588f36" containerName="glance-log" containerID="cri-o://73931af753973ecbc7d6a2fb76ee1d915b3c46b1c26854e339e6a75ae6eb271d" gracePeriod=30 Oct 14 15:33:32 crc kubenswrapper[4945]: I1014 15:33:32.976376 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-conductor-0"] Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.676947 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.742912 4945 generic.go:334] "Generic (PLEG): container finished" podID="bb5c5b1d-4161-435d-8ebe-61016248b180" containerID="2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344" exitCode=0 Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.743283 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bb5c5b1d-4161-435d-8ebe-61016248b180","Type":"ContainerDied","Data":"2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344"} Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.743310 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bb5c5b1d-4161-435d-8ebe-61016248b180","Type":"ContainerDied","Data":"cb518d21296871c66dfe57407a3a0c0d8cada70748427f63b2d35c4b8cad2a7b"} Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.743327 4945 scope.go:117] "RemoveContainer" containerID="2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.743456 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.769423 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b18d41ba-3cf6-454f-ae01-bc9f35df1210","Type":"ContainerStarted","Data":"20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88"} Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.779623 4945 generic.go:334] "Generic (PLEG): container finished" podID="aa12f709-e635-4685-a892-4c65ba588f36" containerID="73931af753973ecbc7d6a2fb76ee1d915b3c46b1c26854e339e6a75ae6eb271d" exitCode=143 Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.779679 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aa12f709-e635-4685-a892-4c65ba588f36","Type":"ContainerDied","Data":"73931af753973ecbc7d6a2fb76ee1d915b3c46b1c26854e339e6a75ae6eb271d"} Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.781973 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"042f4fc3-10ec-450e-b3cf-77a11d71a8b7","Type":"ContainerStarted","Data":"06714dbd05b553ff7840edfe17da91b0c0aef45ca485424eded1711e9aa3994b"} Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.782039 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"042f4fc3-10ec-450e-b3cf-77a11d71a8b7","Type":"ContainerStarted","Data":"7f9a51d4966ef43f1af505efef1183acbf9918990f7a346504b88f84aaf1be6a"} Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.800567 4945 scope.go:117] "RemoveContainer" containerID="166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.814138 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-scripts\") pod \"bb5c5b1d-4161-435d-8ebe-61016248b180\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.814231 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4sfg\" (UniqueName: \"kubernetes.io/projected/bb5c5b1d-4161-435d-8ebe-61016248b180-kube-api-access-l4sfg\") pod \"bb5c5b1d-4161-435d-8ebe-61016248b180\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.814361 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb5c5b1d-4161-435d-8ebe-61016248b180-logs\") pod \"bb5c5b1d-4161-435d-8ebe-61016248b180\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.814406 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-config-data\") pod \"bb5c5b1d-4161-435d-8ebe-61016248b180\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.814446 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb5c5b1d-4161-435d-8ebe-61016248b180-httpd-run\") pod \"bb5c5b1d-4161-435d-8ebe-61016248b180\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.814469 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-internal-tls-certs\") pod \"bb5c5b1d-4161-435d-8ebe-61016248b180\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.814505 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-combined-ca-bundle\") pod \"bb5c5b1d-4161-435d-8ebe-61016248b180\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.814554 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"bb5c5b1d-4161-435d-8ebe-61016248b180\" (UID: \"bb5c5b1d-4161-435d-8ebe-61016248b180\") " Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.817136 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb5c5b1d-4161-435d-8ebe-61016248b180-logs" (OuterVolumeSpecName: "logs") pod "bb5c5b1d-4161-435d-8ebe-61016248b180" (UID: "bb5c5b1d-4161-435d-8ebe-61016248b180"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.821051 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-scripts" (OuterVolumeSpecName: "scripts") pod "bb5c5b1d-4161-435d-8ebe-61016248b180" (UID: "bb5c5b1d-4161-435d-8ebe-61016248b180"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.821180 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb5c5b1d-4161-435d-8ebe-61016248b180-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bb5c5b1d-4161-435d-8ebe-61016248b180" (UID: "bb5c5b1d-4161-435d-8ebe-61016248b180"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.824071 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb5c5b1d-4161-435d-8ebe-61016248b180-kube-api-access-l4sfg" (OuterVolumeSpecName: "kube-api-access-l4sfg") pod "bb5c5b1d-4161-435d-8ebe-61016248b180" (UID: "bb5c5b1d-4161-435d-8ebe-61016248b180"). InnerVolumeSpecName "kube-api-access-l4sfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.831364 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "bb5c5b1d-4161-435d-8ebe-61016248b180" (UID: "bb5c5b1d-4161-435d-8ebe-61016248b180"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.865498 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb5c5b1d-4161-435d-8ebe-61016248b180" (UID: "bb5c5b1d-4161-435d-8ebe-61016248b180"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.900795 4945 scope.go:117] "RemoveContainer" containerID="2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344" Oct 14 15:33:33 crc kubenswrapper[4945]: E1014 15:33:33.901575 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344\": container with ID starting with 2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344 not found: ID does not exist" containerID="2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.901611 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344"} err="failed to get container status \"2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344\": rpc error: code = NotFound desc = could not find container \"2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344\": container with ID starting with 2dcb2c3c068e948178ef41f826c55a08bf23d99f06a2afa03df7b938df41a344 not found: ID does not exist" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.901645 4945 scope.go:117] "RemoveContainer" containerID="166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547" Oct 14 15:33:33 crc kubenswrapper[4945]: E1014 15:33:33.902043 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547\": container with ID starting with 166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547 not found: ID does not exist" containerID="166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.902080 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547"} err="failed to get container status \"166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547\": rpc error: code = NotFound desc = could not find container \"166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547\": container with ID starting with 166779870d89d74ffb6582178b9a84874e1abc0c9fd46db465567e9ff7ad0547 not found: ID does not exist" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.917912 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4sfg\" (UniqueName: \"kubernetes.io/projected/bb5c5b1d-4161-435d-8ebe-61016248b180-kube-api-access-l4sfg\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.917943 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb5c5b1d-4161-435d-8ebe-61016248b180-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.917956 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb5c5b1d-4161-435d-8ebe-61016248b180-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.917969 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.917999 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.918013 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.923686 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-config-data" (OuterVolumeSpecName: "config-data") pod "bb5c5b1d-4161-435d-8ebe-61016248b180" (UID: "bb5c5b1d-4161-435d-8ebe-61016248b180"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.945258 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bb5c5b1d-4161-435d-8ebe-61016248b180" (UID: "bb5c5b1d-4161-435d-8ebe-61016248b180"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:33 crc kubenswrapper[4945]: I1014 15:33:33.949544 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.023668 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.023710 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb5c5b1d-4161-435d-8ebe-61016248b180-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.023725 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.023784 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-bf87b4676-qpt54"] Oct 14 15:33:34 crc kubenswrapper[4945]: E1014 15:33:34.040109 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5c5b1d-4161-435d-8ebe-61016248b180" containerName="glance-httpd" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.040152 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5c5b1d-4161-435d-8ebe-61016248b180" containerName="glance-httpd" Oct 14 15:33:34 crc kubenswrapper[4945]: E1014 15:33:34.040175 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5c5b1d-4161-435d-8ebe-61016248b180" containerName="glance-log" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.040182 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5c5b1d-4161-435d-8ebe-61016248b180" containerName="glance-log" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.040378 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb5c5b1d-4161-435d-8ebe-61016248b180" containerName="glance-log" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.040392 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb5c5b1d-4161-435d-8ebe-61016248b180" containerName="glance-httpd" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.041636 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.049676 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-internal-svc" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.050635 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-public-svc" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.078227 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-bf87b4676-qpt54"] Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.125129 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-etc-podinfo\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.125397 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-config-data\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.125501 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-config-data-custom\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.125587 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-scripts\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.125667 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-combined-ca-bundle\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.125742 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-internal-tls-certs\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.125824 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-config-data-merged\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.125908 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-public-tls-certs\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.126050 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-logs\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.126118 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rpkc\" (UniqueName: \"kubernetes.io/projected/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-kube-api-access-9rpkc\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.145686 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.151830 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.173423 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.174972 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.179579 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.180501 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.190504 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231202 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rpkc\" (UniqueName: \"kubernetes.io/projected/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-kube-api-access-9rpkc\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231246 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr8nm\" (UniqueName: \"kubernetes.io/projected/314804da-3d6d-44cd-a9cf-0f014d690fe9-kube-api-access-tr8nm\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231280 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/314804da-3d6d-44cd-a9cf-0f014d690fe9-logs\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231312 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/314804da-3d6d-44cd-a9cf-0f014d690fe9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231339 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-etc-podinfo\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231367 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-config-data\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231384 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231404 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/314804da-3d6d-44cd-a9cf-0f014d690fe9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231422 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/314804da-3d6d-44cd-a9cf-0f014d690fe9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231447 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-config-data-custom\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231463 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/314804da-3d6d-44cd-a9cf-0f014d690fe9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231494 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-scripts\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231513 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/314804da-3d6d-44cd-a9cf-0f014d690fe9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231538 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-combined-ca-bundle\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231576 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-internal-tls-certs\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231607 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-config-data-merged\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231630 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-public-tls-certs\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.231697 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-logs\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.232149 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-logs\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.236073 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-config-data-merged\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.241890 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-config-data-custom\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.245206 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-combined-ca-bundle\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.245861 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-config-data\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.248806 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-public-tls-certs\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.252670 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-internal-tls-certs\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.253347 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rpkc\" (UniqueName: \"kubernetes.io/projected/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-kube-api-access-9rpkc\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.266647 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-etc-podinfo\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.270955 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f9f45b7-0e91-4ecf-b38f-49bb07f40868-scripts\") pod \"ironic-bf87b4676-qpt54\" (UID: \"5f9f45b7-0e91-4ecf-b38f-49bb07f40868\") " pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.336060 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.336190 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/314804da-3d6d-44cd-a9cf-0f014d690fe9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.336238 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/314804da-3d6d-44cd-a9cf-0f014d690fe9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.336303 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/314804da-3d6d-44cd-a9cf-0f014d690fe9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.336403 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/314804da-3d6d-44cd-a9cf-0f014d690fe9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.336557 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.361364 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr8nm\" (UniqueName: \"kubernetes.io/projected/314804da-3d6d-44cd-a9cf-0f014d690fe9-kube-api-access-tr8nm\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.361460 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/314804da-3d6d-44cd-a9cf-0f014d690fe9-logs\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.361525 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/314804da-3d6d-44cd-a9cf-0f014d690fe9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.367050 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/314804da-3d6d-44cd-a9cf-0f014d690fe9-logs\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.371761 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/314804da-3d6d-44cd-a9cf-0f014d690fe9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.375009 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/314804da-3d6d-44cd-a9cf-0f014d690fe9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.375351 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/314804da-3d6d-44cd-a9cf-0f014d690fe9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.376272 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/314804da-3d6d-44cd-a9cf-0f014d690fe9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.379009 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/314804da-3d6d-44cd-a9cf-0f014d690fe9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.381427 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-8hrxc" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.390303 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr8nm\" (UniqueName: \"kubernetes.io/projected/314804da-3d6d-44cd-a9cf-0f014d690fe9-kube-api-access-tr8nm\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.428522 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.445691 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"314804da-3d6d-44cd-a9cf-0f014d690fe9\") " pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.474296 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gv5g8\" (UniqueName: \"kubernetes.io/projected/92ed91ce-91c5-46ff-8e6c-959b9fa96552-kube-api-access-gv5g8\") pod \"92ed91ce-91c5-46ff-8e6c-959b9fa96552\" (UID: \"92ed91ce-91c5-46ff-8e6c-959b9fa96552\") " Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.487321 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92ed91ce-91c5-46ff-8e6c-959b9fa96552-kube-api-access-gv5g8" (OuterVolumeSpecName: "kube-api-access-gv5g8") pod "92ed91ce-91c5-46ff-8e6c-959b9fa96552" (UID: "92ed91ce-91c5-46ff-8e6c-959b9fa96552"). InnerVolumeSpecName "kube-api-access-gv5g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.577254 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gv5g8\" (UniqueName: \"kubernetes.io/projected/92ed91ce-91c5-46ff-8e6c-959b9fa96552-kube-api-access-gv5g8\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.693528 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.777779 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb5c5b1d-4161-435d-8ebe-61016248b180" path="/var/lib/kubelet/pods/bb5c5b1d-4161-435d-8ebe-61016248b180/volumes" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.796014 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-8hrxc" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.796012 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-8hrxc" event={"ID":"92ed91ce-91c5-46ff-8e6c-959b9fa96552","Type":"ContainerDied","Data":"75bb16074604883801ff2dbf80d9e4fa9f2ba408e1208f872acec9d0f682963b"} Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.796421 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75bb16074604883801ff2dbf80d9e4fa9f2ba408e1208f872acec9d0f682963b" Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.808602 4945 generic.go:334] "Generic (PLEG): container finished" podID="042f4fc3-10ec-450e-b3cf-77a11d71a8b7" containerID="06714dbd05b553ff7840edfe17da91b0c0aef45ca485424eded1711e9aa3994b" exitCode=0 Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.808672 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"042f4fc3-10ec-450e-b3cf-77a11d71a8b7","Type":"ContainerDied","Data":"06714dbd05b553ff7840edfe17da91b0c0aef45ca485424eded1711e9aa3994b"} Oct 14 15:33:34 crc kubenswrapper[4945]: I1014 15:33:34.816340 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b18d41ba-3cf6-454f-ae01-bc9f35df1210","Type":"ContainerStarted","Data":"57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81"} Oct 14 15:33:35 crc kubenswrapper[4945]: I1014 15:33:35.003951 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-bf87b4676-qpt54"] Oct 14 15:33:35 crc kubenswrapper[4945]: I1014 15:33:35.368970 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 15:33:35 crc kubenswrapper[4945]: I1014 15:33:35.827742 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bf87b4676-qpt54" event={"ID":"5f9f45b7-0e91-4ecf-b38f-49bb07f40868","Type":"ContainerStarted","Data":"be72399ac61158aa62942c2c6ddcc927a753f39fb34e6c39d3059e85293b88eb"} Oct 14 15:33:36 crc kubenswrapper[4945]: W1014 15:33:36.014811 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod314804da_3d6d_44cd_a9cf_0f014d690fe9.slice/crio-db3b319425c49d429ee4a9d559041dc3014fd2b4ae328ba8632d91fec3f63157 WatchSource:0}: Error finding container db3b319425c49d429ee4a9d559041dc3014fd2b4ae328ba8632d91fec3f63157: Status 404 returned error can't find the container with id db3b319425c49d429ee4a9d559041dc3014fd2b4ae328ba8632d91fec3f63157 Oct 14 15:33:36 crc kubenswrapper[4945]: I1014 15:33:36.459641 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7cb8f5594f-b9dt5" Oct 14 15:33:36 crc kubenswrapper[4945]: I1014 15:33:36.540664 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-794d4664d6-ns65b"] Oct 14 15:33:36 crc kubenswrapper[4945]: I1014 15:33:36.541426 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-794d4664d6-ns65b" podUID="786749d9-049d-4391-8220-8eb69462f5d1" containerName="neutron-api" containerID="cri-o://a73b648136829109d27da9b0af403d40da63e0076e21a11215b81e2a1c1b0fd4" gracePeriod=30 Oct 14 15:33:36 crc kubenswrapper[4945]: I1014 15:33:36.541478 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-794d4664d6-ns65b" podUID="786749d9-049d-4391-8220-8eb69462f5d1" containerName="neutron-httpd" containerID="cri-o://d70f25ff17a42303a4c62305888fa4fe15b5e474589473276093b7e05c215022" gracePeriod=30 Oct 14 15:33:36 crc kubenswrapper[4945]: I1014 15:33:36.865226 4945 generic.go:334] "Generic (PLEG): container finished" podID="aa12f709-e635-4685-a892-4c65ba588f36" containerID="b241e8b17c3ea5dd6f7019383c566e00ac84c1af7a799b95ea7ad241b3f1a0a8" exitCode=0 Oct 14 15:33:36 crc kubenswrapper[4945]: I1014 15:33:36.865317 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aa12f709-e635-4685-a892-4c65ba588f36","Type":"ContainerDied","Data":"b241e8b17c3ea5dd6f7019383c566e00ac84c1af7a799b95ea7ad241b3f1a0a8"} Oct 14 15:33:36 crc kubenswrapper[4945]: I1014 15:33:36.884036 4945 generic.go:334] "Generic (PLEG): container finished" podID="786749d9-049d-4391-8220-8eb69462f5d1" containerID="d70f25ff17a42303a4c62305888fa4fe15b5e474589473276093b7e05c215022" exitCode=0 Oct 14 15:33:36 crc kubenswrapper[4945]: I1014 15:33:36.884106 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-794d4664d6-ns65b" event={"ID":"786749d9-049d-4391-8220-8eb69462f5d1","Type":"ContainerDied","Data":"d70f25ff17a42303a4c62305888fa4fe15b5e474589473276093b7e05c215022"} Oct 14 15:33:36 crc kubenswrapper[4945]: I1014 15:33:36.892684 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"314804da-3d6d-44cd-a9cf-0f014d690fe9","Type":"ContainerStarted","Data":"db3b319425c49d429ee4a9d559041dc3014fd2b4ae328ba8632d91fec3f63157"} Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.048310 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.134424 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-combined-ca-bundle\") pod \"aa12f709-e635-4685-a892-4c65ba588f36\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.134501 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa12f709-e635-4685-a892-4c65ba588f36-httpd-run\") pod \"aa12f709-e635-4685-a892-4c65ba588f36\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.134572 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa12f709-e635-4685-a892-4c65ba588f36-logs\") pod \"aa12f709-e635-4685-a892-4c65ba588f36\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.134598 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4rmc\" (UniqueName: \"kubernetes.io/projected/aa12f709-e635-4685-a892-4c65ba588f36-kube-api-access-f4rmc\") pod \"aa12f709-e635-4685-a892-4c65ba588f36\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.134626 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-public-tls-certs\") pod \"aa12f709-e635-4685-a892-4c65ba588f36\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.134669 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-config-data\") pod \"aa12f709-e635-4685-a892-4c65ba588f36\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.134758 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-scripts\") pod \"aa12f709-e635-4685-a892-4c65ba588f36\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.134776 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"aa12f709-e635-4685-a892-4c65ba588f36\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.137507 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa12f709-e635-4685-a892-4c65ba588f36-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "aa12f709-e635-4685-a892-4c65ba588f36" (UID: "aa12f709-e635-4685-a892-4c65ba588f36"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.137844 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa12f709-e635-4685-a892-4c65ba588f36-logs" (OuterVolumeSpecName: "logs") pod "aa12f709-e635-4685-a892-4c65ba588f36" (UID: "aa12f709-e635-4685-a892-4c65ba588f36"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.147599 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "aa12f709-e635-4685-a892-4c65ba588f36" (UID: "aa12f709-e635-4685-a892-4c65ba588f36"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.148071 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa12f709-e635-4685-a892-4c65ba588f36-kube-api-access-f4rmc" (OuterVolumeSpecName: "kube-api-access-f4rmc") pod "aa12f709-e635-4685-a892-4c65ba588f36" (UID: "aa12f709-e635-4685-a892-4c65ba588f36"). InnerVolumeSpecName "kube-api-access-f4rmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.176699 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-scripts" (OuterVolumeSpecName: "scripts") pod "aa12f709-e635-4685-a892-4c65ba588f36" (UID: "aa12f709-e635-4685-a892-4c65ba588f36"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.200315 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa12f709-e635-4685-a892-4c65ba588f36" (UID: "aa12f709-e635-4685-a892-4c65ba588f36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.237263 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-config-data" (OuterVolumeSpecName: "config-data") pod "aa12f709-e635-4685-a892-4c65ba588f36" (UID: "aa12f709-e635-4685-a892-4c65ba588f36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.237965 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-config-data\") pod \"aa12f709-e635-4685-a892-4c65ba588f36\" (UID: \"aa12f709-e635-4685-a892-4c65ba588f36\") " Oct 14 15:33:37 crc kubenswrapper[4945]: W1014 15:33:37.238131 4945 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/aa12f709-e635-4685-a892-4c65ba588f36/volumes/kubernetes.io~secret/config-data Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.238148 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-config-data" (OuterVolumeSpecName: "config-data") pod "aa12f709-e635-4685-a892-4c65ba588f36" (UID: "aa12f709-e635-4685-a892-4c65ba588f36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.238735 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.238763 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa12f709-e635-4685-a892-4c65ba588f36-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.238777 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa12f709-e635-4685-a892-4c65ba588f36-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.238788 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4rmc\" (UniqueName: \"kubernetes.io/projected/aa12f709-e635-4685-a892-4c65ba588f36-kube-api-access-f4rmc\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.238802 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.238817 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.238852 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.241814 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "aa12f709-e635-4685-a892-4c65ba588f36" (UID: "aa12f709-e635-4685-a892-4c65ba588f36"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.321229 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.340677 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.340731 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa12f709-e635-4685-a892-4c65ba588f36-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.384923 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.385269 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-86b568845c-dgcjh" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.907842 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" event={"ID":"11506640-cea7-421c-878c-c4b449df880f","Type":"ContainerStarted","Data":"3ce0d470a0c5675fce085a3b000b0e22d5f2706116d70f5eb9c46ec47617e77e"} Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.907905 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.914199 4945 generic.go:334] "Generic (PLEG): container finished" podID="5f9f45b7-0e91-4ecf-b38f-49bb07f40868" containerID="fdb4aab0a00294564ee41eb18e81c4acf50b2a6e21a40d5e67706071692c6a99" exitCode=0 Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.914257 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bf87b4676-qpt54" event={"ID":"5f9f45b7-0e91-4ecf-b38f-49bb07f40868","Type":"ContainerDied","Data":"fdb4aab0a00294564ee41eb18e81c4acf50b2a6e21a40d5e67706071692c6a99"} Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.935324 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b18d41ba-3cf6-454f-ae01-bc9f35df1210","Type":"ContainerStarted","Data":"e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2"} Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.935452 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.935453 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="ceilometer-central-agent" containerID="cri-o://2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9" gracePeriod=30 Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.935486 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="sg-core" containerID="cri-o://57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81" gracePeriod=30 Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.935482 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="proxy-httpd" containerID="cri-o://e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2" gracePeriod=30 Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.935547 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="ceilometer-notification-agent" containerID="cri-o://20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88" gracePeriod=30 Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.939533 4945 generic.go:334] "Generic (PLEG): container finished" podID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerID="f5b4af75350f41bc1d6ed07972ce697303024171cca45c3d4620afcd8b8540c4" exitCode=0 Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.939744 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-78479b8b7d-pdhlw" event={"ID":"5a183ba7-f9df-45a6-8f41-43e4d16eebdd","Type":"ContainerDied","Data":"f5b4af75350f41bc1d6ed07972ce697303024171cca45c3d4620afcd8b8540c4"} Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.949044 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aa12f709-e635-4685-a892-4c65ba588f36","Type":"ContainerDied","Data":"56c00879666ebf1855e58b9c558528904444d29b6e4a601c5816af8327509e90"} Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.949083 4945 scope.go:117] "RemoveContainer" containerID="b241e8b17c3ea5dd6f7019383c566e00ac84c1af7a799b95ea7ad241b3f1a0a8" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.949240 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.951374 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" podStartSLOduration=3.185720222 podStartE2EDuration="7.951346347s" podCreationTimestamp="2025-10-14 15:33:30 +0000 UTC" firstStartedPulling="2025-10-14 15:33:31.929372419 +0000 UTC m=+1141.913420777" lastFinishedPulling="2025-10-14 15:33:36.694998534 +0000 UTC m=+1146.679046902" observedRunningTime="2025-10-14 15:33:37.926719953 +0000 UTC m=+1147.910768321" watchObservedRunningTime="2025-10-14 15:33:37.951346347 +0000 UTC m=+1147.935394735" Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.977069 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"314804da-3d6d-44cd-a9cf-0f014d690fe9","Type":"ContainerStarted","Data":"3a89db091d1f1dbef3ae8c26f012e12dcec05a86a96e6ceb89dc9e605fa6e86f"} Oct 14 15:33:37 crc kubenswrapper[4945]: I1014 15:33:37.996208 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.860503197 podStartE2EDuration="7.99619113s" podCreationTimestamp="2025-10-14 15:33:30 +0000 UTC" firstStartedPulling="2025-10-14 15:33:31.55927442 +0000 UTC m=+1141.543322788" lastFinishedPulling="2025-10-14 15:33:36.694962353 +0000 UTC m=+1146.679010721" observedRunningTime="2025-10-14 15:33:37.989372235 +0000 UTC m=+1147.973420623" watchObservedRunningTime="2025-10-14 15:33:37.99619113 +0000 UTC m=+1147.980239498" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.015203 4945 scope.go:117] "RemoveContainer" containerID="73931af753973ecbc7d6a2fb76ee1d915b3c46b1c26854e339e6a75ae6eb271d" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.075316 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.090888 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.101092 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:33:38 crc kubenswrapper[4945]: E1014 15:33:38.101594 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ed91ce-91c5-46ff-8e6c-959b9fa96552" containerName="mariadb-database-create" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.101611 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ed91ce-91c5-46ff-8e6c-959b9fa96552" containerName="mariadb-database-create" Oct 14 15:33:38 crc kubenswrapper[4945]: E1014 15:33:38.101642 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa12f709-e635-4685-a892-4c65ba588f36" containerName="glance-httpd" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.101651 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa12f709-e635-4685-a892-4c65ba588f36" containerName="glance-httpd" Oct 14 15:33:38 crc kubenswrapper[4945]: E1014 15:33:38.101671 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa12f709-e635-4685-a892-4c65ba588f36" containerName="glance-log" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.101679 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa12f709-e635-4685-a892-4c65ba588f36" containerName="glance-log" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.101929 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="92ed91ce-91c5-46ff-8e6c-959b9fa96552" containerName="mariadb-database-create" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.101955 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa12f709-e635-4685-a892-4c65ba588f36" containerName="glance-httpd" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.101976 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa12f709-e635-4685-a892-4c65ba588f36" containerName="glance-log" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.110573 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: E1014 15:33:38.111479 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f9f45b7_0e91_4ecf_b38f_49bb07f40868.slice/crio-conmon-fdb4aab0a00294564ee41eb18e81c4acf50b2a6e21a40d5e67706071692c6a99.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a183ba7_f9df_45a6_8f41_43e4d16eebdd.slice/crio-conmon-f5b4af75350f41bc1d6ed07972ce697303024171cca45c3d4620afcd8b8540c4.scope\": RecentStats: unable to find data in memory cache]" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.121590 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.123334 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.125862 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.166112 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/57cbaab0-b5ce-4581-af58-87e3e3ec679c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.166376 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/57cbaab0-b5ce-4581-af58-87e3e3ec679c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.166406 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.166431 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57cbaab0-b5ce-4581-af58-87e3e3ec679c-config-data\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.166455 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57cbaab0-b5ce-4581-af58-87e3e3ec679c-scripts\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.166521 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvkb2\" (UniqueName: \"kubernetes.io/projected/57cbaab0-b5ce-4581-af58-87e3e3ec679c-kube-api-access-dvkb2\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.166578 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57cbaab0-b5ce-4581-af58-87e3e3ec679c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.166615 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57cbaab0-b5ce-4581-af58-87e3e3ec679c-logs\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.267961 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvkb2\" (UniqueName: \"kubernetes.io/projected/57cbaab0-b5ce-4581-af58-87e3e3ec679c-kube-api-access-dvkb2\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.268359 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57cbaab0-b5ce-4581-af58-87e3e3ec679c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.270646 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57cbaab0-b5ce-4581-af58-87e3e3ec679c-logs\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.270954 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/57cbaab0-b5ce-4581-af58-87e3e3ec679c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.271161 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/57cbaab0-b5ce-4581-af58-87e3e3ec679c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.278208 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.278343 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57cbaab0-b5ce-4581-af58-87e3e3ec679c-config-data\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.278423 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57cbaab0-b5ce-4581-af58-87e3e3ec679c-scripts\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.272222 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57cbaab0-b5ce-4581-af58-87e3e3ec679c-logs\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.278533 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.273927 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/57cbaab0-b5ce-4581-af58-87e3e3ec679c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.277721 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57cbaab0-b5ce-4581-af58-87e3e3ec679c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.284169 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57cbaab0-b5ce-4581-af58-87e3e3ec679c-config-data\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.287559 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57cbaab0-b5ce-4581-af58-87e3e3ec679c-scripts\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.287731 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/57cbaab0-b5ce-4581-af58-87e3e3ec679c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.288951 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvkb2\" (UniqueName: \"kubernetes.io/projected/57cbaab0-b5ce-4581-af58-87e3e3ec679c-kube-api-access-dvkb2\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.323945 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"57cbaab0-b5ce-4581-af58-87e3e3ec679c\") " pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.465808 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.803638 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa12f709-e635-4685-a892-4c65ba588f36" path="/var/lib/kubelet/pods/aa12f709-e635-4685-a892-4c65ba588f36/volumes" Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.996304 4945 generic.go:334] "Generic (PLEG): container finished" podID="786749d9-049d-4391-8220-8eb69462f5d1" containerID="a73b648136829109d27da9b0af403d40da63e0076e21a11215b81e2a1c1b0fd4" exitCode=0 Oct 14 15:33:38 crc kubenswrapper[4945]: I1014 15:33:38.996367 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-794d4664d6-ns65b" event={"ID":"786749d9-049d-4391-8220-8eb69462f5d1","Type":"ContainerDied","Data":"a73b648136829109d27da9b0af403d40da63e0076e21a11215b81e2a1c1b0fd4"} Oct 14 15:33:39 crc kubenswrapper[4945]: I1014 15:33:39.000220 4945 generic.go:334] "Generic (PLEG): container finished" podID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerID="e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2" exitCode=0 Oct 14 15:33:39 crc kubenswrapper[4945]: I1014 15:33:39.000256 4945 generic.go:334] "Generic (PLEG): container finished" podID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerID="57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81" exitCode=2 Oct 14 15:33:39 crc kubenswrapper[4945]: I1014 15:33:39.000270 4945 generic.go:334] "Generic (PLEG): container finished" podID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerID="20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88" exitCode=0 Oct 14 15:33:39 crc kubenswrapper[4945]: I1014 15:33:39.000293 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b18d41ba-3cf6-454f-ae01-bc9f35df1210","Type":"ContainerDied","Data":"e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2"} Oct 14 15:33:39 crc kubenswrapper[4945]: I1014 15:33:39.000322 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b18d41ba-3cf6-454f-ae01-bc9f35df1210","Type":"ContainerDied","Data":"57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81"} Oct 14 15:33:39 crc kubenswrapper[4945]: I1014 15:33:39.000332 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b18d41ba-3cf6-454f-ae01-bc9f35df1210","Type":"ContainerDied","Data":"20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88"} Oct 14 15:33:39 crc kubenswrapper[4945]: I1014 15:33:39.134563 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 15:33:39 crc kubenswrapper[4945]: W1014 15:33:39.144391 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57cbaab0_b5ce_4581_af58_87e3e3ec679c.slice/crio-3a73365cf933ba4c9b67d6d599a7cb87a66f5ecf7fea41577f47d36262c7bc40 WatchSource:0}: Error finding container 3a73365cf933ba4c9b67d6d599a7cb87a66f5ecf7fea41577f47d36262c7bc40: Status 404 returned error can't find the container with id 3a73365cf933ba4c9b67d6d599a7cb87a66f5ecf7fea41577f47d36262c7bc40 Oct 14 15:33:39 crc kubenswrapper[4945]: I1014 15:33:39.904897 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.016102 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-794d4664d6-ns65b" event={"ID":"786749d9-049d-4391-8220-8eb69462f5d1","Type":"ContainerDied","Data":"baf7d5fc59e2fb444f9738369733286209c618b8b4d512f631199eae47f201a7"} Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.016454 4945 scope.go:117] "RemoveContainer" containerID="d70f25ff17a42303a4c62305888fa4fe15b5e474589473276093b7e05c215022" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.016597 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-794d4664d6-ns65b" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.022326 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-combined-ca-bundle\") pod \"786749d9-049d-4391-8220-8eb69462f5d1\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.022378 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-httpd-config\") pod \"786749d9-049d-4391-8220-8eb69462f5d1\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.022446 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4fcl\" (UniqueName: \"kubernetes.io/projected/786749d9-049d-4391-8220-8eb69462f5d1-kube-api-access-p4fcl\") pod \"786749d9-049d-4391-8220-8eb69462f5d1\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.023386 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-config\") pod \"786749d9-049d-4391-8220-8eb69462f5d1\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.023447 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-ovndb-tls-certs\") pod \"786749d9-049d-4391-8220-8eb69462f5d1\" (UID: \"786749d9-049d-4391-8220-8eb69462f5d1\") " Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.030071 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"314804da-3d6d-44cd-a9cf-0f014d690fe9","Type":"ContainerStarted","Data":"7c6b814ca6288b70e657bcecda10a1037a79a5d10cf5494b521f09873c35511f"} Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.033094 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "786749d9-049d-4391-8220-8eb69462f5d1" (UID: "786749d9-049d-4391-8220-8eb69462f5d1"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.057973 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bf87b4676-qpt54" event={"ID":"5f9f45b7-0e91-4ecf-b38f-49bb07f40868","Type":"ContainerStarted","Data":"f86a815082daa51ff7322770a567ad455a7ef74e1a75cc4f4e7b268f2b4f3dda"} Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.058024 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bf87b4676-qpt54" event={"ID":"5f9f45b7-0e91-4ecf-b38f-49bb07f40868","Type":"ContainerStarted","Data":"1b3a8140ed2a8b15b7a4092aeb69e381a9f039fbac86b2ba3ec4a6da603ee6d9"} Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.058058 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.059071 4945 scope.go:117] "RemoveContainer" containerID="a73b648136829109d27da9b0af403d40da63e0076e21a11215b81e2a1c1b0fd4" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.059388 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/786749d9-049d-4391-8220-8eb69462f5d1-kube-api-access-p4fcl" (OuterVolumeSpecName: "kube-api-access-p4fcl") pod "786749d9-049d-4391-8220-8eb69462f5d1" (UID: "786749d9-049d-4391-8220-8eb69462f5d1"). InnerVolumeSpecName "kube-api-access-p4fcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.068943 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.068920792 podStartE2EDuration="6.068920792s" podCreationTimestamp="2025-10-14 15:33:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:33:40.054272643 +0000 UTC m=+1150.038321011" watchObservedRunningTime="2025-10-14 15:33:40.068920792 +0000 UTC m=+1150.052969160" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.074422 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"57cbaab0-b5ce-4581-af58-87e3e3ec679c","Type":"ContainerStarted","Data":"3a73365cf933ba4c9b67d6d599a7cb87a66f5ecf7fea41577f47d36262c7bc40"} Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.087433 4945 generic.go:334] "Generic (PLEG): container finished" podID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerID="7783a506aa052fa817cbd2ab866b5bac20da7549ba6276c2c64b1fbd5b33cc0c" exitCode=1 Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.087489 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-78479b8b7d-pdhlw" event={"ID":"5a183ba7-f9df-45a6-8f41-43e4d16eebdd","Type":"ContainerDied","Data":"7783a506aa052fa817cbd2ab866b5bac20da7549ba6276c2c64b1fbd5b33cc0c"} Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.087521 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-78479b8b7d-pdhlw" event={"ID":"5a183ba7-f9df-45a6-8f41-43e4d16eebdd","Type":"ContainerStarted","Data":"571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0"} Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.088116 4945 scope.go:117] "RemoveContainer" containerID="7783a506aa052fa817cbd2ab866b5bac20da7549ba6276c2c64b1fbd5b33cc0c" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.091073 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-bf87b4676-qpt54" podStartSLOduration=5.410090192 podStartE2EDuration="7.091053385s" podCreationTimestamp="2025-10-14 15:33:33 +0000 UTC" firstStartedPulling="2025-10-14 15:33:35.013089634 +0000 UTC m=+1144.997138002" lastFinishedPulling="2025-10-14 15:33:36.694052827 +0000 UTC m=+1146.678101195" observedRunningTime="2025-10-14 15:33:40.09054261 +0000 UTC m=+1150.074591008" watchObservedRunningTime="2025-10-14 15:33:40.091053385 +0000 UTC m=+1150.075101763" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.125935 4945 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.125968 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4fcl\" (UniqueName: \"kubernetes.io/projected/786749d9-049d-4391-8220-8eb69462f5d1-kube-api-access-p4fcl\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.145730 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "786749d9-049d-4391-8220-8eb69462f5d1" (UID: "786749d9-049d-4391-8220-8eb69462f5d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.185236 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "786749d9-049d-4391-8220-8eb69462f5d1" (UID: "786749d9-049d-4391-8220-8eb69462f5d1"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.189031 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-config" (OuterVolumeSpecName: "config") pod "786749d9-049d-4391-8220-8eb69462f5d1" (UID: "786749d9-049d-4391-8220-8eb69462f5d1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.231403 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.231441 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.231456 4945 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/786749d9-049d-4391-8220-8eb69462f5d1-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.358391 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-794d4664d6-ns65b"] Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.368806 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-794d4664d6-ns65b"] Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.776060 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="786749d9-049d-4391-8220-8eb69462f5d1" path="/var/lib/kubelet/pods/786749d9-049d-4391-8220-8eb69462f5d1/volumes" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.840759 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-c7c3-account-create-pz2vd"] Oct 14 15:33:40 crc kubenswrapper[4945]: E1014 15:33:40.841603 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786749d9-049d-4391-8220-8eb69462f5d1" containerName="neutron-api" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.841651 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="786749d9-049d-4391-8220-8eb69462f5d1" containerName="neutron-api" Oct 14 15:33:40 crc kubenswrapper[4945]: E1014 15:33:40.841679 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786749d9-049d-4391-8220-8eb69462f5d1" containerName="neutron-httpd" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.841687 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="786749d9-049d-4391-8220-8eb69462f5d1" containerName="neutron-httpd" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.842100 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="786749d9-049d-4391-8220-8eb69462f5d1" containerName="neutron-api" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.842160 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="786749d9-049d-4391-8220-8eb69462f5d1" containerName="neutron-httpd" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.843330 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-c7c3-account-create-pz2vd" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.853246 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-c7c3-account-create-pz2vd"] Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.853526 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-db-secret" Oct 14 15:33:40 crc kubenswrapper[4945]: I1014 15:33:40.945798 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7br8\" (UniqueName: \"kubernetes.io/projected/6d9430f4-426a-40f2-8c6d-c9ce794d48a3-kube-api-access-x7br8\") pod \"ironic-inspector-c7c3-account-create-pz2vd\" (UID: \"6d9430f4-426a-40f2-8c6d-c9ce794d48a3\") " pod="openstack/ironic-inspector-c7c3-account-create-pz2vd" Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.047101 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7br8\" (UniqueName: \"kubernetes.io/projected/6d9430f4-426a-40f2-8c6d-c9ce794d48a3-kube-api-access-x7br8\") pod \"ironic-inspector-c7c3-account-create-pz2vd\" (UID: \"6d9430f4-426a-40f2-8c6d-c9ce794d48a3\") " pod="openstack/ironic-inspector-c7c3-account-create-pz2vd" Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.068862 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7br8\" (UniqueName: \"kubernetes.io/projected/6d9430f4-426a-40f2-8c6d-c9ce794d48a3-kube-api-access-x7br8\") pod \"ironic-inspector-c7c3-account-create-pz2vd\" (UID: \"6d9430f4-426a-40f2-8c6d-c9ce794d48a3\") " pod="openstack/ironic-inspector-c7c3-account-create-pz2vd" Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.108693 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"57cbaab0-b5ce-4581-af58-87e3e3ec679c","Type":"ContainerStarted","Data":"d950eb114961c508917ee0d1a8a98533ee28a4dab769b85faf63430a64624b1f"} Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.108744 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"57cbaab0-b5ce-4581-af58-87e3e3ec679c","Type":"ContainerStarted","Data":"c072e03e8345acae7d63feb854445377b571798d0149528a2879edcbc0c4c30d"} Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.114218 4945 generic.go:334] "Generic (PLEG): container finished" podID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerID="6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd" exitCode=1 Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.114288 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-78479b8b7d-pdhlw" event={"ID":"5a183ba7-f9df-45a6-8f41-43e4d16eebdd","Type":"ContainerDied","Data":"6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd"} Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.114328 4945 scope.go:117] "RemoveContainer" containerID="7783a506aa052fa817cbd2ab866b5bac20da7549ba6276c2c64b1fbd5b33cc0c" Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.115107 4945 scope.go:117] "RemoveContainer" containerID="6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd" Oct 14 15:33:41 crc kubenswrapper[4945]: E1014 15:33:41.115347 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-api pod=ironic-78479b8b7d-pdhlw_openstack(5a183ba7-f9df-45a6-8f41-43e4d16eebdd)\"" pod="openstack/ironic-78479b8b7d-pdhlw" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.118680 4945 generic.go:334] "Generic (PLEG): container finished" podID="11506640-cea7-421c-878c-c4b449df880f" containerID="3ce0d470a0c5675fce085a3b000b0e22d5f2706116d70f5eb9c46ec47617e77e" exitCode=1 Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.119439 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" event={"ID":"11506640-cea7-421c-878c-c4b449df880f","Type":"ContainerDied","Data":"3ce0d470a0c5675fce085a3b000b0e22d5f2706116d70f5eb9c46ec47617e77e"} Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.119707 4945 scope.go:117] "RemoveContainer" containerID="3ce0d470a0c5675fce085a3b000b0e22d5f2706116d70f5eb9c46ec47617e77e" Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.129717 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.12969654 podStartE2EDuration="3.12969654s" podCreationTimestamp="2025-10-14 15:33:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:33:41.125380216 +0000 UTC m=+1151.109428584" watchObservedRunningTime="2025-10-14 15:33:41.12969654 +0000 UTC m=+1151.113744918" Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.196056 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-c7c3-account-create-pz2vd" Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.267445 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.441601 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:41 crc kubenswrapper[4945]: I1014 15:33:41.441645 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:42 crc kubenswrapper[4945]: I1014 15:33:42.181232 4945 scope.go:117] "RemoveContainer" containerID="6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd" Oct 14 15:33:42 crc kubenswrapper[4945]: E1014 15:33:42.181618 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-api pod=ironic-78479b8b7d-pdhlw_openstack(5a183ba7-f9df-45a6-8f41-43e4d16eebdd)\"" pod="openstack/ironic-78479b8b7d-pdhlw" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" Oct 14 15:33:42 crc kubenswrapper[4945]: I1014 15:33:42.538494 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-c7c3-account-create-pz2vd"] Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.189924 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" event={"ID":"11506640-cea7-421c-878c-c4b449df880f","Type":"ContainerStarted","Data":"40d15dc11255c9246509b77850ad41e5fafbb3b18e210f748158b2b666b15891"} Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.190226 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.193082 4945 generic.go:334] "Generic (PLEG): container finished" podID="6d9430f4-426a-40f2-8c6d-c9ce794d48a3" containerID="ddd5cd09cc3e25634705a0e23971e72527bc2353839ffbf5969d4617e8e393b9" exitCode=0 Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.193180 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-c7c3-account-create-pz2vd" event={"ID":"6d9430f4-426a-40f2-8c6d-c9ce794d48a3","Type":"ContainerDied","Data":"ddd5cd09cc3e25634705a0e23971e72527bc2353839ffbf5969d4617e8e393b9"} Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.193208 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-c7c3-account-create-pz2vd" event={"ID":"6d9430f4-426a-40f2-8c6d-c9ce794d48a3","Type":"ContainerStarted","Data":"b00ef7cbe986a433e4ec174d8e6783bbc07c4d017d4e884062865dbaa0dd77d4"} Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.196575 4945 scope.go:117] "RemoveContainer" containerID="6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd" Oct 14 15:33:43 crc kubenswrapper[4945]: E1014 15:33:43.197100 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-api pod=ironic-78479b8b7d-pdhlw_openstack(5a183ba7-f9df-45a6-8f41-43e4d16eebdd)\"" pod="openstack/ironic-78479b8b7d-pdhlw" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.718592 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-bf87b4676-qpt54" Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.783044 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-78479b8b7d-pdhlw"] Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.862313 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.917212 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b18d41ba-3cf6-454f-ae01-bc9f35df1210-run-httpd\") pod \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.917265 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-combined-ca-bundle\") pod \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.917314 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-sg-core-conf-yaml\") pod \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.917390 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kckfx\" (UniqueName: \"kubernetes.io/projected/b18d41ba-3cf6-454f-ae01-bc9f35df1210-kube-api-access-kckfx\") pod \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.917440 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b18d41ba-3cf6-454f-ae01-bc9f35df1210-log-httpd\") pod \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.917468 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-scripts\") pod \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.917529 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-config-data\") pod \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\" (UID: \"b18d41ba-3cf6-454f-ae01-bc9f35df1210\") " Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.918246 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b18d41ba-3cf6-454f-ae01-bc9f35df1210-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b18d41ba-3cf6-454f-ae01-bc9f35df1210" (UID: "b18d41ba-3cf6-454f-ae01-bc9f35df1210"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.918697 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b18d41ba-3cf6-454f-ae01-bc9f35df1210-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b18d41ba-3cf6-454f-ae01-bc9f35df1210" (UID: "b18d41ba-3cf6-454f-ae01-bc9f35df1210"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.923919 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b18d41ba-3cf6-454f-ae01-bc9f35df1210-kube-api-access-kckfx" (OuterVolumeSpecName: "kube-api-access-kckfx") pod "b18d41ba-3cf6-454f-ae01-bc9f35df1210" (UID: "b18d41ba-3cf6-454f-ae01-bc9f35df1210"). InnerVolumeSpecName "kube-api-access-kckfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.936487 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-scripts" (OuterVolumeSpecName: "scripts") pod "b18d41ba-3cf6-454f-ae01-bc9f35df1210" (UID: "b18d41ba-3cf6-454f-ae01-bc9f35df1210"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:43 crc kubenswrapper[4945]: I1014 15:33:43.964834 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b18d41ba-3cf6-454f-ae01-bc9f35df1210" (UID: "b18d41ba-3cf6-454f-ae01-bc9f35df1210"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.019130 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b18d41ba-3cf6-454f-ae01-bc9f35df1210-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.019159 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.019168 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b18d41ba-3cf6-454f-ae01-bc9f35df1210-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.019175 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.019185 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kckfx\" (UniqueName: \"kubernetes.io/projected/b18d41ba-3cf6-454f-ae01-bc9f35df1210-kube-api-access-kckfx\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.041148 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-config-data" (OuterVolumeSpecName: "config-data") pod "b18d41ba-3cf6-454f-ae01-bc9f35df1210" (UID: "b18d41ba-3cf6-454f-ae01-bc9f35df1210"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.046668 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b18d41ba-3cf6-454f-ae01-bc9f35df1210" (UID: "b18d41ba-3cf6-454f-ae01-bc9f35df1210"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.121086 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.121116 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b18d41ba-3cf6-454f-ae01-bc9f35df1210-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.206062 4945 generic.go:334] "Generic (PLEG): container finished" podID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerID="2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9" exitCode=0 Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.206140 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.206197 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b18d41ba-3cf6-454f-ae01-bc9f35df1210","Type":"ContainerDied","Data":"2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9"} Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.206235 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b18d41ba-3cf6-454f-ae01-bc9f35df1210","Type":"ContainerDied","Data":"b332e70687114937a0ab9a38b755ee566e1b736b8cdaf87324e18fd3ddc794eb"} Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.206256 4945 scope.go:117] "RemoveContainer" containerID="e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.206567 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ironic-78479b8b7d-pdhlw" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerName="ironic-api-log" containerID="cri-o://571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0" gracePeriod=60 Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.263085 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.269805 4945 scope.go:117] "RemoveContainer" containerID="57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.273043 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.280621 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:44 crc kubenswrapper[4945]: E1014 15:33:44.281057 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="sg-core" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.281077 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="sg-core" Oct 14 15:33:44 crc kubenswrapper[4945]: E1014 15:33:44.281101 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="ceilometer-notification-agent" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.281111 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="ceilometer-notification-agent" Oct 14 15:33:44 crc kubenswrapper[4945]: E1014 15:33:44.281133 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="ceilometer-central-agent" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.281141 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="ceilometer-central-agent" Oct 14 15:33:44 crc kubenswrapper[4945]: E1014 15:33:44.281180 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="proxy-httpd" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.281188 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="proxy-httpd" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.281393 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="proxy-httpd" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.281420 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="sg-core" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.281450 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="ceilometer-notification-agent" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.281465 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" containerName="ceilometer-central-agent" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.283125 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.285319 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.286572 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.290306 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.327565 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.327633 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-run-httpd\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.327698 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpxk5\" (UniqueName: \"kubernetes.io/projected/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-kube-api-access-fpxk5\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.327731 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-config-data\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.327778 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-log-httpd\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.327809 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.327897 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-scripts\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.355110 4945 scope.go:117] "RemoveContainer" containerID="20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.400608 4945 scope.go:117] "RemoveContainer" containerID="2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.429452 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpxk5\" (UniqueName: \"kubernetes.io/projected/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-kube-api-access-fpxk5\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.429510 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-config-data\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.429567 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-log-httpd\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.429592 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.429672 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-scripts\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.429734 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.429773 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-run-httpd\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.430238 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-run-httpd\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.430647 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-log-httpd\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.441989 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.446422 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.446446 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-scripts\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.446837 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-config-data\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.450654 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpxk5\" (UniqueName: \"kubernetes.io/projected/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-kube-api-access-fpxk5\") pod \"ceilometer-0\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.462729 4945 scope.go:117] "RemoveContainer" containerID="e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2" Oct 14 15:33:44 crc kubenswrapper[4945]: E1014 15:33:44.463198 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2\": container with ID starting with e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2 not found: ID does not exist" containerID="e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.463238 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2"} err="failed to get container status \"e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2\": rpc error: code = NotFound desc = could not find container \"e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2\": container with ID starting with e246c364befef24067edaf4a76d1df774131cf7e42444dbd71c0a548744aaff2 not found: ID does not exist" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.463268 4945 scope.go:117] "RemoveContainer" containerID="57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81" Oct 14 15:33:44 crc kubenswrapper[4945]: E1014 15:33:44.463668 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81\": container with ID starting with 57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81 not found: ID does not exist" containerID="57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.463699 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81"} err="failed to get container status \"57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81\": rpc error: code = NotFound desc = could not find container \"57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81\": container with ID starting with 57397a5d5ef9e210dde3528f2ae12f1d32d84a8766c1aba1bf01a1d241284c81 not found: ID does not exist" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.463725 4945 scope.go:117] "RemoveContainer" containerID="20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88" Oct 14 15:33:44 crc kubenswrapper[4945]: E1014 15:33:44.464058 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88\": container with ID starting with 20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88 not found: ID does not exist" containerID="20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.464102 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88"} err="failed to get container status \"20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88\": rpc error: code = NotFound desc = could not find container \"20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88\": container with ID starting with 20ce14ac46f94c479c8335ee9a0e61228481ed4f7861c62289b1658a22e0ac88 not found: ID does not exist" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.464119 4945 scope.go:117] "RemoveContainer" containerID="2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9" Oct 14 15:33:44 crc kubenswrapper[4945]: E1014 15:33:44.464599 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9\": container with ID starting with 2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9 not found: ID does not exist" containerID="2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.464626 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9"} err="failed to get container status \"2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9\": rpc error: code = NotFound desc = could not find container \"2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9\": container with ID starting with 2f6ea2ab0d6da17d83df26658117b6d607940e921688c797281cbc192ca05bf9 not found: ID does not exist" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.672256 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-c7c3-account-create-pz2vd" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.676764 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.695158 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.695215 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.744213 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.751945 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.782642 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b18d41ba-3cf6-454f-ae01-bc9f35df1210" path="/var/lib/kubelet/pods/b18d41ba-3cf6-454f-ae01-bc9f35df1210/volumes" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.794850 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.836336 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7br8\" (UniqueName: \"kubernetes.io/projected/6d9430f4-426a-40f2-8c6d-c9ce794d48a3-kube-api-access-x7br8\") pod \"6d9430f4-426a-40f2-8c6d-c9ce794d48a3\" (UID: \"6d9430f4-426a-40f2-8c6d-c9ce794d48a3\") " Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.844975 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d9430f4-426a-40f2-8c6d-c9ce794d48a3-kube-api-access-x7br8" (OuterVolumeSpecName: "kube-api-access-x7br8") pod "6d9430f4-426a-40f2-8c6d-c9ce794d48a3" (UID: "6d9430f4-426a-40f2-8c6d-c9ce794d48a3"). InnerVolumeSpecName "kube-api-access-x7br8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.938015 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbgtp\" (UniqueName: \"kubernetes.io/projected/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-kube-api-access-sbgtp\") pod \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.938658 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data\") pod \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.938680 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data-custom\") pod \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.938769 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-combined-ca-bundle\") pod \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.938847 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-logs\") pod \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.939041 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data-merged\") pod \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.939096 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-scripts\") pod \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.939129 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-etc-podinfo\") pod \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\" (UID: \"5a183ba7-f9df-45a6-8f41-43e4d16eebdd\") " Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.939679 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-logs" (OuterVolumeSpecName: "logs") pod "5a183ba7-f9df-45a6-8f41-43e4d16eebdd" (UID: "5a183ba7-f9df-45a6-8f41-43e4d16eebdd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.940392 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.940408 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7br8\" (UniqueName: \"kubernetes.io/projected/6d9430f4-426a-40f2-8c6d-c9ce794d48a3-kube-api-access-x7br8\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.940948 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "5a183ba7-f9df-45a6-8f41-43e4d16eebdd" (UID: "5a183ba7-f9df-45a6-8f41-43e4d16eebdd"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.943739 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5a183ba7-f9df-45a6-8f41-43e4d16eebdd" (UID: "5a183ba7-f9df-45a6-8f41-43e4d16eebdd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.943737 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-kube-api-access-sbgtp" (OuterVolumeSpecName: "kube-api-access-sbgtp") pod "5a183ba7-f9df-45a6-8f41-43e4d16eebdd" (UID: "5a183ba7-f9df-45a6-8f41-43e4d16eebdd"). InnerVolumeSpecName "kube-api-access-sbgtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.945746 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-scripts" (OuterVolumeSpecName: "scripts") pod "5a183ba7-f9df-45a6-8f41-43e4d16eebdd" (UID: "5a183ba7-f9df-45a6-8f41-43e4d16eebdd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.947111 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "5a183ba7-f9df-45a6-8f41-43e4d16eebdd" (UID: "5a183ba7-f9df-45a6-8f41-43e4d16eebdd"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.973631 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data" (OuterVolumeSpecName: "config-data") pod "5a183ba7-f9df-45a6-8f41-43e4d16eebdd" (UID: "5a183ba7-f9df-45a6-8f41-43e4d16eebdd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:44 crc kubenswrapper[4945]: I1014 15:33:44.993624 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a183ba7-f9df-45a6-8f41-43e4d16eebdd" (UID: "5a183ba7-f9df-45a6-8f41-43e4d16eebdd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.042021 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.042059 4945 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-etc-podinfo\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.042073 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbgtp\" (UniqueName: \"kubernetes.io/projected/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-kube-api-access-sbgtp\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.042082 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.042092 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.042102 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.042111 4945 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5a183ba7-f9df-45a6-8f41-43e4d16eebdd-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.192099 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:45 crc kubenswrapper[4945]: W1014 15:33:45.202032 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod847a6bf4_b43c_45f5_9b11_5d0cad77da3e.slice/crio-db80b64b4082fb28c5f9add9ff3a840b8bcd0ba1f6d407417378171e719330a8 WatchSource:0}: Error finding container db80b64b4082fb28c5f9add9ff3a840b8bcd0ba1f6d407417378171e719330a8: Status 404 returned error can't find the container with id db80b64b4082fb28c5f9add9ff3a840b8bcd0ba1f6d407417378171e719330a8 Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.219290 4945 generic.go:334] "Generic (PLEG): container finished" podID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerID="571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0" exitCode=143 Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.219354 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-78479b8b7d-pdhlw" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.219391 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-78479b8b7d-pdhlw" event={"ID":"5a183ba7-f9df-45a6-8f41-43e4d16eebdd","Type":"ContainerDied","Data":"571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0"} Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.219447 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-78479b8b7d-pdhlw" event={"ID":"5a183ba7-f9df-45a6-8f41-43e4d16eebdd","Type":"ContainerDied","Data":"afa08e7038da40a958ca7bc580c1ab7f1f231dad3ea9bc61e6979463e772698a"} Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.219467 4945 scope.go:117] "RemoveContainer" containerID="6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.221993 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"847a6bf4-b43c-45f5-9b11-5d0cad77da3e","Type":"ContainerStarted","Data":"db80b64b4082fb28c5f9add9ff3a840b8bcd0ba1f6d407417378171e719330a8"} Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.224314 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-c7c3-account-create-pz2vd" event={"ID":"6d9430f4-426a-40f2-8c6d-c9ce794d48a3","Type":"ContainerDied","Data":"b00ef7cbe986a433e4ec174d8e6783bbc07c4d017d4e884062865dbaa0dd77d4"} Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.224350 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b00ef7cbe986a433e4ec174d8e6783bbc07c4d017d4e884062865dbaa0dd77d4" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.224407 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-c7c3-account-create-pz2vd" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.229251 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.229283 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.279111 4945 scope.go:117] "RemoveContainer" containerID="571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.324653 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-78479b8b7d-pdhlw"] Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.330599 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-78479b8b7d-pdhlw"] Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.334077 4945 scope.go:117] "RemoveContainer" containerID="f5b4af75350f41bc1d6ed07972ce697303024171cca45c3d4620afcd8b8540c4" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.357431 4945 scope.go:117] "RemoveContainer" containerID="6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd" Oct 14 15:33:45 crc kubenswrapper[4945]: E1014 15:33:45.357944 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd\": container with ID starting with 6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd not found: ID does not exist" containerID="6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.357975 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd"} err="failed to get container status \"6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd\": rpc error: code = NotFound desc = could not find container \"6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd\": container with ID starting with 6b80321760888b34219bc429f81618ac1da8ad15fc3042e34b983733936bd9dd not found: ID does not exist" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.357996 4945 scope.go:117] "RemoveContainer" containerID="571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0" Oct 14 15:33:45 crc kubenswrapper[4945]: E1014 15:33:45.358450 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0\": container with ID starting with 571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0 not found: ID does not exist" containerID="571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.358476 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0"} err="failed to get container status \"571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0\": rpc error: code = NotFound desc = could not find container \"571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0\": container with ID starting with 571f1634276739766a7e076a2cc814497006a80b776858e7dca821269aa795b0 not found: ID does not exist" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.358492 4945 scope.go:117] "RemoveContainer" containerID="f5b4af75350f41bc1d6ed07972ce697303024171cca45c3d4620afcd8b8540c4" Oct 14 15:33:45 crc kubenswrapper[4945]: E1014 15:33:45.358922 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5b4af75350f41bc1d6ed07972ce697303024171cca45c3d4620afcd8b8540c4\": container with ID starting with f5b4af75350f41bc1d6ed07972ce697303024171cca45c3d4620afcd8b8540c4 not found: ID does not exist" containerID="f5b4af75350f41bc1d6ed07972ce697303024171cca45c3d4620afcd8b8540c4" Oct 14 15:33:45 crc kubenswrapper[4945]: I1014 15:33:45.359021 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5b4af75350f41bc1d6ed07972ce697303024171cca45c3d4620afcd8b8540c4"} err="failed to get container status \"f5b4af75350f41bc1d6ed07972ce697303024171cca45c3d4620afcd8b8540c4\": rpc error: code = NotFound desc = could not find container \"f5b4af75350f41bc1d6ed07972ce697303024171cca45c3d4620afcd8b8540c4\": container with ID starting with f5b4af75350f41bc1d6ed07972ce697303024171cca45c3d4620afcd8b8540c4 not found: ID does not exist" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.238161 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"847a6bf4-b43c-45f5-9b11-5d0cad77da3e","Type":"ContainerStarted","Data":"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda"} Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.239816 4945 generic.go:334] "Generic (PLEG): container finished" podID="11506640-cea7-421c-878c-c4b449df880f" containerID="40d15dc11255c9246509b77850ad41e5fafbb3b18e210f748158b2b666b15891" exitCode=1 Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.239866 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" event={"ID":"11506640-cea7-421c-878c-c4b449df880f","Type":"ContainerDied","Data":"40d15dc11255c9246509b77850ad41e5fafbb3b18e210f748158b2b666b15891"} Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.239912 4945 scope.go:117] "RemoveContainer" containerID="3ce0d470a0c5675fce085a3b000b0e22d5f2706116d70f5eb9c46ec47617e77e" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.240563 4945 scope.go:117] "RemoveContainer" containerID="40d15dc11255c9246509b77850ad41e5fafbb3b18e210f748158b2b666b15891" Oct 14 15:33:46 crc kubenswrapper[4945]: E1014 15:33:46.240898 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-648bd876cc-srkfr_openstack(11506640-cea7-421c-878c-c4b449df880f)\"" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" podUID="11506640-cea7-421c-878c-c4b449df880f" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.266777 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.554831 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-xqd5f"] Oct 14 15:33:46 crc kubenswrapper[4945]: E1014 15:33:46.555291 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerName="ironic-api" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.555313 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerName="ironic-api" Oct 14 15:33:46 crc kubenswrapper[4945]: E1014 15:33:46.555333 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerName="ironic-api-log" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.555341 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerName="ironic-api-log" Oct 14 15:33:46 crc kubenswrapper[4945]: E1014 15:33:46.555359 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d9430f4-426a-40f2-8c6d-c9ce794d48a3" containerName="mariadb-account-create" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.555368 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d9430f4-426a-40f2-8c6d-c9ce794d48a3" containerName="mariadb-account-create" Oct 14 15:33:46 crc kubenswrapper[4945]: E1014 15:33:46.555391 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerName="init" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.555399 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerName="init" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.555604 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerName="ironic-api-log" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.555616 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d9430f4-426a-40f2-8c6d-c9ce794d48a3" containerName="mariadb-account-create" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.555635 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerName="ironic-api" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.557290 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xqd5f" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.563935 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xqd5f"] Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.600699 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbdtk\" (UniqueName: \"kubernetes.io/projected/874d6996-5c8a-4be4-b3e0-ed7a102cf8ba-kube-api-access-xbdtk\") pod \"nova-api-db-create-xqd5f\" (UID: \"874d6996-5c8a-4be4-b3e0-ed7a102cf8ba\") " pod="openstack/nova-api-db-create-xqd5f" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.662152 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-k49f8"] Oct 14 15:33:46 crc kubenswrapper[4945]: E1014 15:33:46.662534 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerName="ironic-api" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.662552 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerName="ironic-api" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.662738 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" containerName="ironic-api" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.663435 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-k49f8" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.671357 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-k49f8"] Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.703075 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8q4w\" (UniqueName: \"kubernetes.io/projected/ae872229-e7d9-4fcd-b22d-650f80e8f18a-kube-api-access-q8q4w\") pod \"nova-cell0-db-create-k49f8\" (UID: \"ae872229-e7d9-4fcd-b22d-650f80e8f18a\") " pod="openstack/nova-cell0-db-create-k49f8" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.703355 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbdtk\" (UniqueName: \"kubernetes.io/projected/874d6996-5c8a-4be4-b3e0-ed7a102cf8ba-kube-api-access-xbdtk\") pod \"nova-api-db-create-xqd5f\" (UID: \"874d6996-5c8a-4be4-b3e0-ed7a102cf8ba\") " pod="openstack/nova-api-db-create-xqd5f" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.721000 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbdtk\" (UniqueName: \"kubernetes.io/projected/874d6996-5c8a-4be4-b3e0-ed7a102cf8ba-kube-api-access-xbdtk\") pod \"nova-api-db-create-xqd5f\" (UID: \"874d6996-5c8a-4be4-b3e0-ed7a102cf8ba\") " pod="openstack/nova-api-db-create-xqd5f" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.774753 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a183ba7-f9df-45a6-8f41-43e4d16eebdd" path="/var/lib/kubelet/pods/5a183ba7-f9df-45a6-8f41-43e4d16eebdd/volumes" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.804243 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8q4w\" (UniqueName: \"kubernetes.io/projected/ae872229-e7d9-4fcd-b22d-650f80e8f18a-kube-api-access-q8q4w\") pod \"nova-cell0-db-create-k49f8\" (UID: \"ae872229-e7d9-4fcd-b22d-650f80e8f18a\") " pod="openstack/nova-cell0-db-create-k49f8" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.827926 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8q4w\" (UniqueName: \"kubernetes.io/projected/ae872229-e7d9-4fcd-b22d-650f80e8f18a-kube-api-access-q8q4w\") pod \"nova-cell0-db-create-k49f8\" (UID: \"ae872229-e7d9-4fcd-b22d-650f80e8f18a\") " pod="openstack/nova-cell0-db-create-k49f8" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.876170 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xqd5f" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.879930 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-zn84r"] Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.881610 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-zn84r" Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.899584 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-zn84r"] Oct 14 15:33:46 crc kubenswrapper[4945]: I1014 15:33:46.983161 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-k49f8" Oct 14 15:33:47 crc kubenswrapper[4945]: I1014 15:33:47.020672 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnstk\" (UniqueName: \"kubernetes.io/projected/5d684cba-7b02-453e-8fc0-f2607daa9951-kube-api-access-cnstk\") pod \"nova-cell1-db-create-zn84r\" (UID: \"5d684cba-7b02-453e-8fc0-f2607daa9951\") " pod="openstack/nova-cell1-db-create-zn84r" Oct 14 15:33:47 crc kubenswrapper[4945]: I1014 15:33:47.122012 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnstk\" (UniqueName: \"kubernetes.io/projected/5d684cba-7b02-453e-8fc0-f2607daa9951-kube-api-access-cnstk\") pod \"nova-cell1-db-create-zn84r\" (UID: \"5d684cba-7b02-453e-8fc0-f2607daa9951\") " pod="openstack/nova-cell1-db-create-zn84r" Oct 14 15:33:47 crc kubenswrapper[4945]: I1014 15:33:47.142769 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnstk\" (UniqueName: \"kubernetes.io/projected/5d684cba-7b02-453e-8fc0-f2607daa9951-kube-api-access-cnstk\") pod \"nova-cell1-db-create-zn84r\" (UID: \"5d684cba-7b02-453e-8fc0-f2607daa9951\") " pod="openstack/nova-cell1-db-create-zn84r" Oct 14 15:33:47 crc kubenswrapper[4945]: I1014 15:33:47.241942 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-zn84r" Oct 14 15:33:47 crc kubenswrapper[4945]: I1014 15:33:47.254275 4945 scope.go:117] "RemoveContainer" containerID="40d15dc11255c9246509b77850ad41e5fafbb3b18e210f748158b2b666b15891" Oct 14 15:33:47 crc kubenswrapper[4945]: E1014 15:33:47.254856 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-648bd876cc-srkfr_openstack(11506640-cea7-421c-878c-c4b449df880f)\"" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" podUID="11506640-cea7-421c-878c-c4b449df880f" Oct 14 15:33:47 crc kubenswrapper[4945]: I1014 15:33:47.258368 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"847a6bf4-b43c-45f5-9b11-5d0cad77da3e","Type":"ContainerStarted","Data":"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257"} Oct 14 15:33:47 crc kubenswrapper[4945]: I1014 15:33:47.276206 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:47 crc kubenswrapper[4945]: I1014 15:33:47.276298 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 15:33:47 crc kubenswrapper[4945]: I1014 15:33:47.293065 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 15:33:47 crc kubenswrapper[4945]: I1014 15:33:47.467097 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xqd5f"] Oct 14 15:33:47 crc kubenswrapper[4945]: I1014 15:33:47.637324 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-k49f8"] Oct 14 15:33:47 crc kubenswrapper[4945]: W1014 15:33:47.642745 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae872229_e7d9_4fcd_b22d_650f80e8f18a.slice/crio-ab01ccd1609a0f05ef201d930ebd0766f1fb1e1a0d94618599d93ae932260187 WatchSource:0}: Error finding container ab01ccd1609a0f05ef201d930ebd0766f1fb1e1a0d94618599d93ae932260187: Status 404 returned error can't find the container with id ab01ccd1609a0f05ef201d930ebd0766f1fb1e1a0d94618599d93ae932260187 Oct 14 15:33:47 crc kubenswrapper[4945]: I1014 15:33:47.774313 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-zn84r"] Oct 14 15:33:47 crc kubenswrapper[4945]: W1014 15:33:47.777808 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d684cba_7b02_453e_8fc0_f2607daa9951.slice/crio-b3ed0dbbe05843393e0f808db17a13e7c67eceac3c7ebbfe7edb64bf838024f4 WatchSource:0}: Error finding container b3ed0dbbe05843393e0f808db17a13e7c67eceac3c7ebbfe7edb64bf838024f4: Status 404 returned error can't find the container with id b3ed0dbbe05843393e0f808db17a13e7c67eceac3c7ebbfe7edb64bf838024f4 Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.271913 4945 generic.go:334] "Generic (PLEG): container finished" podID="874d6996-5c8a-4be4-b3e0-ed7a102cf8ba" containerID="d735d29138108f68908aa7b5a4bcc6e8783d56a23a5935ab20e297cf863d5279" exitCode=0 Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.272090 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xqd5f" event={"ID":"874d6996-5c8a-4be4-b3e0-ed7a102cf8ba","Type":"ContainerDied","Data":"d735d29138108f68908aa7b5a4bcc6e8783d56a23a5935ab20e297cf863d5279"} Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.272300 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xqd5f" event={"ID":"874d6996-5c8a-4be4-b3e0-ed7a102cf8ba","Type":"ContainerStarted","Data":"77e03f7d4bd41cd101be9265a2ed3bad52c47d95f83e59f9c77b2281cecba75a"} Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.273795 4945 generic.go:334] "Generic (PLEG): container finished" podID="ae872229-e7d9-4fcd-b22d-650f80e8f18a" containerID="4d9f58b702d01e9dc96b268b5d5e40d9d58a217ce6c8016419c5bd675cb251e3" exitCode=0 Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.273839 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-k49f8" event={"ID":"ae872229-e7d9-4fcd-b22d-650f80e8f18a","Type":"ContainerDied","Data":"4d9f58b702d01e9dc96b268b5d5e40d9d58a217ce6c8016419c5bd675cb251e3"} Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.273857 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-k49f8" event={"ID":"ae872229-e7d9-4fcd-b22d-650f80e8f18a","Type":"ContainerStarted","Data":"ab01ccd1609a0f05ef201d930ebd0766f1fb1e1a0d94618599d93ae932260187"} Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.275501 4945 generic.go:334] "Generic (PLEG): container finished" podID="5d684cba-7b02-453e-8fc0-f2607daa9951" containerID="1ed91e4e41516acc7e6b8f9ad985d95b34f69a93773e12d062643de0ef1e2d41" exitCode=0 Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.275543 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-zn84r" event={"ID":"5d684cba-7b02-453e-8fc0-f2607daa9951","Type":"ContainerDied","Data":"1ed91e4e41516acc7e6b8f9ad985d95b34f69a93773e12d062643de0ef1e2d41"} Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.275559 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-zn84r" event={"ID":"5d684cba-7b02-453e-8fc0-f2607daa9951","Type":"ContainerStarted","Data":"b3ed0dbbe05843393e0f808db17a13e7c67eceac3c7ebbfe7edb64bf838024f4"} Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.278749 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"847a6bf4-b43c-45f5-9b11-5d0cad77da3e","Type":"ContainerStarted","Data":"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277"} Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.466203 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.467097 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.498669 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 15:33:48 crc kubenswrapper[4945]: I1014 15:33:48.531636 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 15:33:49 crc kubenswrapper[4945]: I1014 15:33:49.290536 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 15:33:49 crc kubenswrapper[4945]: I1014 15:33:49.291388 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 15:33:49 crc kubenswrapper[4945]: I1014 15:33:49.732475 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-zn84r" Oct 14 15:33:49 crc kubenswrapper[4945]: I1014 15:33:49.856024 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xqd5f" Oct 14 15:33:49 crc kubenswrapper[4945]: I1014 15:33:49.862387 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-k49f8" Oct 14 15:33:49 crc kubenswrapper[4945]: I1014 15:33:49.878522 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnstk\" (UniqueName: \"kubernetes.io/projected/5d684cba-7b02-453e-8fc0-f2607daa9951-kube-api-access-cnstk\") pod \"5d684cba-7b02-453e-8fc0-f2607daa9951\" (UID: \"5d684cba-7b02-453e-8fc0-f2607daa9951\") " Oct 14 15:33:49 crc kubenswrapper[4945]: I1014 15:33:49.884048 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d684cba-7b02-453e-8fc0-f2607daa9951-kube-api-access-cnstk" (OuterVolumeSpecName: "kube-api-access-cnstk") pod "5d684cba-7b02-453e-8fc0-f2607daa9951" (UID: "5d684cba-7b02-453e-8fc0-f2607daa9951"). InnerVolumeSpecName "kube-api-access-cnstk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:49 crc kubenswrapper[4945]: I1014 15:33:49.980404 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8q4w\" (UniqueName: \"kubernetes.io/projected/ae872229-e7d9-4fcd-b22d-650f80e8f18a-kube-api-access-q8q4w\") pod \"ae872229-e7d9-4fcd-b22d-650f80e8f18a\" (UID: \"ae872229-e7d9-4fcd-b22d-650f80e8f18a\") " Oct 14 15:33:49 crc kubenswrapper[4945]: I1014 15:33:49.980648 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbdtk\" (UniqueName: \"kubernetes.io/projected/874d6996-5c8a-4be4-b3e0-ed7a102cf8ba-kube-api-access-xbdtk\") pod \"874d6996-5c8a-4be4-b3e0-ed7a102cf8ba\" (UID: \"874d6996-5c8a-4be4-b3e0-ed7a102cf8ba\") " Oct 14 15:33:49 crc kubenswrapper[4945]: I1014 15:33:49.981153 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnstk\" (UniqueName: \"kubernetes.io/projected/5d684cba-7b02-453e-8fc0-f2607daa9951-kube-api-access-cnstk\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:49 crc kubenswrapper[4945]: I1014 15:33:49.983960 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae872229-e7d9-4fcd-b22d-650f80e8f18a-kube-api-access-q8q4w" (OuterVolumeSpecName: "kube-api-access-q8q4w") pod "ae872229-e7d9-4fcd-b22d-650f80e8f18a" (UID: "ae872229-e7d9-4fcd-b22d-650f80e8f18a"). InnerVolumeSpecName "kube-api-access-q8q4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:49 crc kubenswrapper[4945]: I1014 15:33:49.984412 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/874d6996-5c8a-4be4-b3e0-ed7a102cf8ba-kube-api-access-xbdtk" (OuterVolumeSpecName: "kube-api-access-xbdtk") pod "874d6996-5c8a-4be4-b3e0-ed7a102cf8ba" (UID: "874d6996-5c8a-4be4-b3e0-ed7a102cf8ba"). InnerVolumeSpecName "kube-api-access-xbdtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:50 crc kubenswrapper[4945]: I1014 15:33:50.082832 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbdtk\" (UniqueName: \"kubernetes.io/projected/874d6996-5c8a-4be4-b3e0-ed7a102cf8ba-kube-api-access-xbdtk\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:50 crc kubenswrapper[4945]: I1014 15:33:50.082904 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8q4w\" (UniqueName: \"kubernetes.io/projected/ae872229-e7d9-4fcd-b22d-650f80e8f18a-kube-api-access-q8q4w\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:50 crc kubenswrapper[4945]: I1014 15:33:50.304134 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-k49f8" event={"ID":"ae872229-e7d9-4fcd-b22d-650f80e8f18a","Type":"ContainerDied","Data":"ab01ccd1609a0f05ef201d930ebd0766f1fb1e1a0d94618599d93ae932260187"} Oct 14 15:33:50 crc kubenswrapper[4945]: I1014 15:33:50.304208 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab01ccd1609a0f05ef201d930ebd0766f1fb1e1a0d94618599d93ae932260187" Oct 14 15:33:50 crc kubenswrapper[4945]: I1014 15:33:50.304347 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-k49f8" Oct 14 15:33:50 crc kubenswrapper[4945]: I1014 15:33:50.308618 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-zn84r" event={"ID":"5d684cba-7b02-453e-8fc0-f2607daa9951","Type":"ContainerDied","Data":"b3ed0dbbe05843393e0f808db17a13e7c67eceac3c7ebbfe7edb64bf838024f4"} Oct 14 15:33:50 crc kubenswrapper[4945]: I1014 15:33:50.308666 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3ed0dbbe05843393e0f808db17a13e7c67eceac3c7ebbfe7edb64bf838024f4" Oct 14 15:33:50 crc kubenswrapper[4945]: I1014 15:33:50.308726 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-zn84r" Oct 14 15:33:50 crc kubenswrapper[4945]: I1014 15:33:50.313348 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xqd5f" Oct 14 15:33:50 crc kubenswrapper[4945]: I1014 15:33:50.313441 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xqd5f" event={"ID":"874d6996-5c8a-4be4-b3e0-ed7a102cf8ba","Type":"ContainerDied","Data":"77e03f7d4bd41cd101be9265a2ed3bad52c47d95f83e59f9c77b2281cecba75a"} Oct 14 15:33:50 crc kubenswrapper[4945]: I1014 15:33:50.313511 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77e03f7d4bd41cd101be9265a2ed3bad52c47d95f83e59f9c77b2281cecba75a" Oct 14 15:33:51 crc kubenswrapper[4945]: I1014 15:33:51.323401 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 15:33:51 crc kubenswrapper[4945]: I1014 15:33:51.323428 4945 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 15:33:51 crc kubenswrapper[4945]: I1014 15:33:51.805499 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 15:33:51 crc kubenswrapper[4945]: I1014 15:33:51.812321 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 15:33:52 crc kubenswrapper[4945]: I1014 15:33:52.333212 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"847a6bf4-b43c-45f5-9b11-5d0cad77da3e","Type":"ContainerStarted","Data":"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d"} Oct 14 15:33:52 crc kubenswrapper[4945]: I1014 15:33:52.357732 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.118003915 podStartE2EDuration="8.357709104s" podCreationTimestamp="2025-10-14 15:33:44 +0000 UTC" firstStartedPulling="2025-10-14 15:33:45.203894143 +0000 UTC m=+1155.187942511" lastFinishedPulling="2025-10-14 15:33:51.443599332 +0000 UTC m=+1161.427647700" observedRunningTime="2025-10-14 15:33:52.348410748 +0000 UTC m=+1162.332459116" watchObservedRunningTime="2025-10-14 15:33:52.357709104 +0000 UTC m=+1162.341757472" Oct 14 15:33:53 crc kubenswrapper[4945]: I1014 15:33:53.343122 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.136843 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.364061 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="ceilometer-central-agent" containerID="cri-o://f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda" gracePeriod=30 Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.364123 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="proxy-httpd" containerID="cri-o://a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d" gracePeriod=30 Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.364175 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="ceilometer-notification-agent" containerID="cri-o://f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257" gracePeriod=30 Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.364142 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="sg-core" containerID="cri-o://fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277" gracePeriod=30 Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.869265 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-db-sync-t252w"] Oct 14 15:33:55 crc kubenswrapper[4945]: E1014 15:33:55.869620 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874d6996-5c8a-4be4-b3e0-ed7a102cf8ba" containerName="mariadb-database-create" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.869637 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="874d6996-5c8a-4be4-b3e0-ed7a102cf8ba" containerName="mariadb-database-create" Oct 14 15:33:55 crc kubenswrapper[4945]: E1014 15:33:55.869659 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d684cba-7b02-453e-8fc0-f2607daa9951" containerName="mariadb-database-create" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.869665 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d684cba-7b02-453e-8fc0-f2607daa9951" containerName="mariadb-database-create" Oct 14 15:33:55 crc kubenswrapper[4945]: E1014 15:33:55.869691 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae872229-e7d9-4fcd-b22d-650f80e8f18a" containerName="mariadb-database-create" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.869698 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae872229-e7d9-4fcd-b22d-650f80e8f18a" containerName="mariadb-database-create" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.869891 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae872229-e7d9-4fcd-b22d-650f80e8f18a" containerName="mariadb-database-create" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.869905 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d684cba-7b02-453e-8fc0-f2607daa9951" containerName="mariadb-database-create" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.869918 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="874d6996-5c8a-4be4-b3e0-ed7a102cf8ba" containerName="mariadb-database-create" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.870500 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.872542 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.872795 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.891859 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/048f7a68-56de-43d2-8ffb-33acacc4e2ec-var-lib-ironic\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.891990 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-config\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.892024 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-combined-ca-bundle\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.892140 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-scripts\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.892215 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfssn\" (UniqueName: \"kubernetes.io/projected/048f7a68-56de-43d2-8ffb-33acacc4e2ec-kube-api-access-sfssn\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.892279 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/048f7a68-56de-43d2-8ffb-33acacc4e2ec-etc-podinfo\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.892307 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/048f7a68-56de-43d2-8ffb-33acacc4e2ec-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.894572 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-sync-t252w"] Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.994271 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-scripts\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.994803 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfssn\" (UniqueName: \"kubernetes.io/projected/048f7a68-56de-43d2-8ffb-33acacc4e2ec-kube-api-access-sfssn\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.994985 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/048f7a68-56de-43d2-8ffb-33acacc4e2ec-etc-podinfo\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.995014 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/048f7a68-56de-43d2-8ffb-33acacc4e2ec-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.995092 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/048f7a68-56de-43d2-8ffb-33acacc4e2ec-var-lib-ironic\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.995206 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-config\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.995236 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-combined-ca-bundle\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.996102 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/048f7a68-56de-43d2-8ffb-33acacc4e2ec-var-lib-ironic\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:55 crc kubenswrapper[4945]: I1014 15:33:55.998032 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/048f7a68-56de-43d2-8ffb-33acacc4e2ec-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.000452 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/048f7a68-56de-43d2-8ffb-33acacc4e2ec-etc-podinfo\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.001837 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-config\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.008143 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-scripts\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.011852 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-combined-ca-bundle\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.021606 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfssn\" (UniqueName: \"kubernetes.io/projected/048f7a68-56de-43d2-8ffb-33acacc4e2ec-kube-api-access-sfssn\") pod \"ironic-inspector-db-sync-t252w\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.236308 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.241517 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.304328 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-config-data\") pod \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.304765 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-run-httpd\") pod \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.304950 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-log-httpd\") pod \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.305022 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-sg-core-conf-yaml\") pod \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.305086 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-combined-ca-bundle\") pod \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.305116 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpxk5\" (UniqueName: \"kubernetes.io/projected/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-kube-api-access-fpxk5\") pod \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.305148 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-scripts\") pod \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\" (UID: \"847a6bf4-b43c-45f5-9b11-5d0cad77da3e\") " Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.306833 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "847a6bf4-b43c-45f5-9b11-5d0cad77da3e" (UID: "847a6bf4-b43c-45f5-9b11-5d0cad77da3e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.307152 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "847a6bf4-b43c-45f5-9b11-5d0cad77da3e" (UID: "847a6bf4-b43c-45f5-9b11-5d0cad77da3e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.316646 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-scripts" (OuterVolumeSpecName: "scripts") pod "847a6bf4-b43c-45f5-9b11-5d0cad77da3e" (UID: "847a6bf4-b43c-45f5-9b11-5d0cad77da3e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.316818 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-kube-api-access-fpxk5" (OuterVolumeSpecName: "kube-api-access-fpxk5") pod "847a6bf4-b43c-45f5-9b11-5d0cad77da3e" (UID: "847a6bf4-b43c-45f5-9b11-5d0cad77da3e"). InnerVolumeSpecName "kube-api-access-fpxk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.335824 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "847a6bf4-b43c-45f5-9b11-5d0cad77da3e" (UID: "847a6bf4-b43c-45f5-9b11-5d0cad77da3e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.377041 4945 generic.go:334] "Generic (PLEG): container finished" podID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerID="a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d" exitCode=0 Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.377078 4945 generic.go:334] "Generic (PLEG): container finished" podID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerID="fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277" exitCode=2 Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.377087 4945 generic.go:334] "Generic (PLEG): container finished" podID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerID="f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257" exitCode=0 Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.377097 4945 generic.go:334] "Generic (PLEG): container finished" podID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerID="f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda" exitCode=0 Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.377119 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"847a6bf4-b43c-45f5-9b11-5d0cad77da3e","Type":"ContainerDied","Data":"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d"} Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.377149 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"847a6bf4-b43c-45f5-9b11-5d0cad77da3e","Type":"ContainerDied","Data":"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277"} Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.377164 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"847a6bf4-b43c-45f5-9b11-5d0cad77da3e","Type":"ContainerDied","Data":"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257"} Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.377175 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"847a6bf4-b43c-45f5-9b11-5d0cad77da3e","Type":"ContainerDied","Data":"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda"} Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.377185 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"847a6bf4-b43c-45f5-9b11-5d0cad77da3e","Type":"ContainerDied","Data":"db80b64b4082fb28c5f9add9ff3a840b8bcd0ba1f6d407417378171e719330a8"} Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.378059 4945 scope.go:117] "RemoveContainer" containerID="a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.378253 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.406523 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "847a6bf4-b43c-45f5-9b11-5d0cad77da3e" (UID: "847a6bf4-b43c-45f5-9b11-5d0cad77da3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.407418 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.407442 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.407504 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.407520 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpxk5\" (UniqueName: \"kubernetes.io/projected/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-kube-api-access-fpxk5\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.407532 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.407543 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.415618 4945 scope.go:117] "RemoveContainer" containerID="fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.437665 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-config-data" (OuterVolumeSpecName: "config-data") pod "847a6bf4-b43c-45f5-9b11-5d0cad77da3e" (UID: "847a6bf4-b43c-45f5-9b11-5d0cad77da3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.438735 4945 scope.go:117] "RemoveContainer" containerID="f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.464077 4945 scope.go:117] "RemoveContainer" containerID="f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.510606 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/847a6bf4-b43c-45f5-9b11-5d0cad77da3e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.548827 4945 scope.go:117] "RemoveContainer" containerID="a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d" Oct 14 15:33:56 crc kubenswrapper[4945]: E1014 15:33:56.549263 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d\": container with ID starting with a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d not found: ID does not exist" containerID="a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.549293 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d"} err="failed to get container status \"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d\": rpc error: code = NotFound desc = could not find container \"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d\": container with ID starting with a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.549318 4945 scope.go:117] "RemoveContainer" containerID="fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277" Oct 14 15:33:56 crc kubenswrapper[4945]: E1014 15:33:56.549772 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277\": container with ID starting with fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277 not found: ID does not exist" containerID="fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.549799 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277"} err="failed to get container status \"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277\": rpc error: code = NotFound desc = could not find container \"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277\": container with ID starting with fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277 not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.549815 4945 scope.go:117] "RemoveContainer" containerID="f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257" Oct 14 15:33:56 crc kubenswrapper[4945]: E1014 15:33:56.550046 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257\": container with ID starting with f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257 not found: ID does not exist" containerID="f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.550070 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257"} err="failed to get container status \"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257\": rpc error: code = NotFound desc = could not find container \"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257\": container with ID starting with f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257 not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.550086 4945 scope.go:117] "RemoveContainer" containerID="f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda" Oct 14 15:33:56 crc kubenswrapper[4945]: E1014 15:33:56.550310 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda\": container with ID starting with f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda not found: ID does not exist" containerID="f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.550337 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda"} err="failed to get container status \"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda\": rpc error: code = NotFound desc = could not find container \"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda\": container with ID starting with f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.550355 4945 scope.go:117] "RemoveContainer" containerID="a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.550611 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d"} err="failed to get container status \"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d\": rpc error: code = NotFound desc = could not find container \"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d\": container with ID starting with a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.550634 4945 scope.go:117] "RemoveContainer" containerID="fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.550820 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277"} err="failed to get container status \"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277\": rpc error: code = NotFound desc = could not find container \"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277\": container with ID starting with fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277 not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.550845 4945 scope.go:117] "RemoveContainer" containerID="f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.551410 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257"} err="failed to get container status \"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257\": rpc error: code = NotFound desc = could not find container \"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257\": container with ID starting with f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257 not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.551435 4945 scope.go:117] "RemoveContainer" containerID="f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.551695 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda"} err="failed to get container status \"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda\": rpc error: code = NotFound desc = could not find container \"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda\": container with ID starting with f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.551718 4945 scope.go:117] "RemoveContainer" containerID="a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.551948 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d"} err="failed to get container status \"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d\": rpc error: code = NotFound desc = could not find container \"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d\": container with ID starting with a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.551969 4945 scope.go:117] "RemoveContainer" containerID="fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.552160 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277"} err="failed to get container status \"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277\": rpc error: code = NotFound desc = could not find container \"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277\": container with ID starting with fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277 not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.552183 4945 scope.go:117] "RemoveContainer" containerID="f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.552441 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257"} err="failed to get container status \"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257\": rpc error: code = NotFound desc = could not find container \"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257\": container with ID starting with f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257 not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.552466 4945 scope.go:117] "RemoveContainer" containerID="f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.552767 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda"} err="failed to get container status \"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda\": rpc error: code = NotFound desc = could not find container \"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda\": container with ID starting with f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.552788 4945 scope.go:117] "RemoveContainer" containerID="a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.553022 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d"} err="failed to get container status \"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d\": rpc error: code = NotFound desc = could not find container \"a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d\": container with ID starting with a24b742e5a8d87d0bd00a8aca16941799c8f44eb15661df923003f7f2c10ec3d not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.553047 4945 scope.go:117] "RemoveContainer" containerID="fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.553385 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277"} err="failed to get container status \"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277\": rpc error: code = NotFound desc = could not find container \"fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277\": container with ID starting with fad792f4d3229379c7de8dc9eaf52f6b5c1076a1d83dcf70f76f9cc0b40be277 not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.553409 4945 scope.go:117] "RemoveContainer" containerID="f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.553578 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257"} err="failed to get container status \"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257\": rpc error: code = NotFound desc = could not find container \"f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257\": container with ID starting with f014a4ef7c8b62fdab59af23aacc5e4615a8d496669504a06ae195ec49610257 not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.553599 4945 scope.go:117] "RemoveContainer" containerID="f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.553833 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda"} err="failed to get container status \"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda\": rpc error: code = NotFound desc = could not find container \"f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda\": container with ID starting with f92582c9c22747e3c856617e956798e65429934bf36286e42496003c45cf3fda not found: ID does not exist" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.720529 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.728560 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.736765 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-sync-t252w"] Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.744738 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-4571-account-create-r2t9d"] Oct 14 15:33:56 crc kubenswrapper[4945]: E1014 15:33:56.745247 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="sg-core" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.745270 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="sg-core" Oct 14 15:33:56 crc kubenswrapper[4945]: E1014 15:33:56.745285 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="ceilometer-central-agent" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.745293 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="ceilometer-central-agent" Oct 14 15:33:56 crc kubenswrapper[4945]: E1014 15:33:56.745319 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="proxy-httpd" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.745327 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="proxy-httpd" Oct 14 15:33:56 crc kubenswrapper[4945]: E1014 15:33:56.745358 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="ceilometer-notification-agent" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.745366 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="ceilometer-notification-agent" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.745586 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="sg-core" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.745611 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="proxy-httpd" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.745635 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="ceilometer-notification-agent" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.745652 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" containerName="ceilometer-central-agent" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.746442 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-4571-account-create-r2t9d" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.749238 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.757931 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.760701 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.764712 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.764946 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.780312 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="847a6bf4-b43c-45f5-9b11-5d0cad77da3e" path="/var/lib/kubelet/pods/847a6bf4-b43c-45f5-9b11-5d0cad77da3e/volumes" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.781339 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-4571-account-create-r2t9d"] Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.781375 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.818072 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-config-data\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.820080 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.820147 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.820185 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7b358e-6730-43b7-bc2b-a154926a32de-run-httpd\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.820342 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7b358e-6730-43b7-bc2b-a154926a32de-log-httpd\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.820885 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47jgw\" (UniqueName: \"kubernetes.io/projected/d0c8cdd9-c134-4013-b110-388be1ead4db-kube-api-access-47jgw\") pod \"nova-api-4571-account-create-r2t9d\" (UID: \"d0c8cdd9-c134-4013-b110-388be1ead4db\") " pod="openstack/nova-api-4571-account-create-r2t9d" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.821000 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-scripts\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.821061 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74phm\" (UniqueName: \"kubernetes.io/projected/2a7b358e-6730-43b7-bc2b-a154926a32de-kube-api-access-74phm\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.922209 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47jgw\" (UniqueName: \"kubernetes.io/projected/d0c8cdd9-c134-4013-b110-388be1ead4db-kube-api-access-47jgw\") pod \"nova-api-4571-account-create-r2t9d\" (UID: \"d0c8cdd9-c134-4013-b110-388be1ead4db\") " pod="openstack/nova-api-4571-account-create-r2t9d" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.922288 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-scripts\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.922314 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74phm\" (UniqueName: \"kubernetes.io/projected/2a7b358e-6730-43b7-bc2b-a154926a32de-kube-api-access-74phm\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.922368 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-config-data\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.922426 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.922449 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.922466 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7b358e-6730-43b7-bc2b-a154926a32de-run-httpd\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.922511 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7b358e-6730-43b7-bc2b-a154926a32de-log-httpd\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.923001 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7b358e-6730-43b7-bc2b-a154926a32de-log-httpd\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.923619 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7b358e-6730-43b7-bc2b-a154926a32de-run-httpd\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.927350 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.927494 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.927980 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-config-data\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.931535 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-scripts\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.938517 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cdc7-account-create-qqwb7"] Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.940031 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cdc7-account-create-qqwb7" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.942473 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.953234 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47jgw\" (UniqueName: \"kubernetes.io/projected/d0c8cdd9-c134-4013-b110-388be1ead4db-kube-api-access-47jgw\") pod \"nova-api-4571-account-create-r2t9d\" (UID: \"d0c8cdd9-c134-4013-b110-388be1ead4db\") " pod="openstack/nova-api-4571-account-create-r2t9d" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.957589 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74phm\" (UniqueName: \"kubernetes.io/projected/2a7b358e-6730-43b7-bc2b-a154926a32de-kube-api-access-74phm\") pod \"ceilometer-0\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " pod="openstack/ceilometer-0" Oct 14 15:33:56 crc kubenswrapper[4945]: I1014 15:33:56.958493 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cdc7-account-create-qqwb7"] Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.024021 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xtjf\" (UniqueName: \"kubernetes.io/projected/9dbf2952-a679-4420-910d-0ec3b3d6e1c1-kube-api-access-2xtjf\") pod \"nova-cell0-cdc7-account-create-qqwb7\" (UID: \"9dbf2952-a679-4420-910d-0ec3b3d6e1c1\") " pod="openstack/nova-cell0-cdc7-account-create-qqwb7" Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.089964 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-4571-account-create-r2t9d" Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.103190 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.125811 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xtjf\" (UniqueName: \"kubernetes.io/projected/9dbf2952-a679-4420-910d-0ec3b3d6e1c1-kube-api-access-2xtjf\") pod \"nova-cell0-cdc7-account-create-qqwb7\" (UID: \"9dbf2952-a679-4420-910d-0ec3b3d6e1c1\") " pod="openstack/nova-cell0-cdc7-account-create-qqwb7" Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.151169 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xtjf\" (UniqueName: \"kubernetes.io/projected/9dbf2952-a679-4420-910d-0ec3b3d6e1c1-kube-api-access-2xtjf\") pod \"nova-cell0-cdc7-account-create-qqwb7\" (UID: \"9dbf2952-a679-4420-910d-0ec3b3d6e1c1\") " pod="openstack/nova-cell0-cdc7-account-create-qqwb7" Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.167825 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ce31-account-create-425x4"] Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.169213 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ce31-account-create-425x4" Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.172086 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ce31-account-create-425x4"] Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.191342 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.228735 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b2sd\" (UniqueName: \"kubernetes.io/projected/82992e17-467c-46cf-8348-4248a6844b71-kube-api-access-6b2sd\") pod \"nova-cell1-ce31-account-create-425x4\" (UID: \"82992e17-467c-46cf-8348-4248a6844b71\") " pod="openstack/nova-cell1-ce31-account-create-425x4" Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.330529 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b2sd\" (UniqueName: \"kubernetes.io/projected/82992e17-467c-46cf-8348-4248a6844b71-kube-api-access-6b2sd\") pod \"nova-cell1-ce31-account-create-425x4\" (UID: \"82992e17-467c-46cf-8348-4248a6844b71\") " pod="openstack/nova-cell1-ce31-account-create-425x4" Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.347313 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b2sd\" (UniqueName: \"kubernetes.io/projected/82992e17-467c-46cf-8348-4248a6844b71-kube-api-access-6b2sd\") pod \"nova-cell1-ce31-account-create-425x4\" (UID: \"82992e17-467c-46cf-8348-4248a6844b71\") " pod="openstack/nova-cell1-ce31-account-create-425x4" Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.352087 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cdc7-account-create-qqwb7" Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.393933 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-t252w" event={"ID":"048f7a68-56de-43d2-8ffb-33acacc4e2ec","Type":"ContainerStarted","Data":"cf81a0f5fb2f18581f3579f5213835ab144d779285b57e3e090d146b993cdcd9"} Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.514094 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ce31-account-create-425x4" Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.650492 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.704292 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-4571-account-create-r2t9d"] Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.824699 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cdc7-account-create-qqwb7"] Oct 14 15:33:57 crc kubenswrapper[4945]: W1014 15:33:57.825232 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9dbf2952_a679_4420_910d_0ec3b3d6e1c1.slice/crio-3e65ac0714910aa063f8088a7c689835f2521a4be7148d87eafd6d009956c077 WatchSource:0}: Error finding container 3e65ac0714910aa063f8088a7c689835f2521a4be7148d87eafd6d009956c077: Status 404 returned error can't find the container with id 3e65ac0714910aa063f8088a7c689835f2521a4be7148d87eafd6d009956c077 Oct 14 15:33:57 crc kubenswrapper[4945]: I1014 15:33:57.988048 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ce31-account-create-425x4"] Oct 14 15:33:58 crc kubenswrapper[4945]: I1014 15:33:58.404689 4945 generic.go:334] "Generic (PLEG): container finished" podID="82992e17-467c-46cf-8348-4248a6844b71" containerID="2c12bdad5e23d70aad90bbf19c2d017a97c95f3c679426d9a0ff06b25caa0089" exitCode=0 Oct 14 15:33:58 crc kubenswrapper[4945]: I1014 15:33:58.404771 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ce31-account-create-425x4" event={"ID":"82992e17-467c-46cf-8348-4248a6844b71","Type":"ContainerDied","Data":"2c12bdad5e23d70aad90bbf19c2d017a97c95f3c679426d9a0ff06b25caa0089"} Oct 14 15:33:58 crc kubenswrapper[4945]: I1014 15:33:58.405124 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ce31-account-create-425x4" event={"ID":"82992e17-467c-46cf-8348-4248a6844b71","Type":"ContainerStarted","Data":"defd92e1c8e62dbb2dd6aaab2b05e7a143bae7ddcd46ee9cd843d272408ec695"} Oct 14 15:33:58 crc kubenswrapper[4945]: I1014 15:33:58.408692 4945 generic.go:334] "Generic (PLEG): container finished" podID="9dbf2952-a679-4420-910d-0ec3b3d6e1c1" containerID="a041f37d48bb8115b6436162ff26bb850f219513d141dc56dde2b6d342b6a886" exitCode=0 Oct 14 15:33:58 crc kubenswrapper[4945]: I1014 15:33:58.408765 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cdc7-account-create-qqwb7" event={"ID":"9dbf2952-a679-4420-910d-0ec3b3d6e1c1","Type":"ContainerDied","Data":"a041f37d48bb8115b6436162ff26bb850f219513d141dc56dde2b6d342b6a886"} Oct 14 15:33:58 crc kubenswrapper[4945]: I1014 15:33:58.408806 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cdc7-account-create-qqwb7" event={"ID":"9dbf2952-a679-4420-910d-0ec3b3d6e1c1","Type":"ContainerStarted","Data":"3e65ac0714910aa063f8088a7c689835f2521a4be7148d87eafd6d009956c077"} Oct 14 15:33:58 crc kubenswrapper[4945]: I1014 15:33:58.410343 4945 generic.go:334] "Generic (PLEG): container finished" podID="d0c8cdd9-c134-4013-b110-388be1ead4db" containerID="60be43dbea963d67d0a7034c4d6b7f50b01a064be46324c64f49a9d21c998559" exitCode=0 Oct 14 15:33:58 crc kubenswrapper[4945]: I1014 15:33:58.410403 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-4571-account-create-r2t9d" event={"ID":"d0c8cdd9-c134-4013-b110-388be1ead4db","Type":"ContainerDied","Data":"60be43dbea963d67d0a7034c4d6b7f50b01a064be46324c64f49a9d21c998559"} Oct 14 15:33:58 crc kubenswrapper[4945]: I1014 15:33:58.410422 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-4571-account-create-r2t9d" event={"ID":"d0c8cdd9-c134-4013-b110-388be1ead4db","Type":"ContainerStarted","Data":"1b8e736c08adcc6d2cd420c6197ddf1391ba7adb1507b2ed40ecb7b3b9385a7e"} Oct 14 15:33:58 crc kubenswrapper[4945]: I1014 15:33:58.411815 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7b358e-6730-43b7-bc2b-a154926a32de","Type":"ContainerStarted","Data":"5f5ad7834f6e37bd63f4a76a3714f09174f167cdb441d1374f580b32694f2771"} Oct 14 15:33:58 crc kubenswrapper[4945]: I1014 15:33:58.411836 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7b358e-6730-43b7-bc2b-a154926a32de","Type":"ContainerStarted","Data":"693c94bd3c6022c1b90f23674544d526844c0fbded319528569ee9a68df0eebc"} Oct 14 15:33:59 crc kubenswrapper[4945]: I1014 15:33:59.424960 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7b358e-6730-43b7-bc2b-a154926a32de","Type":"ContainerStarted","Data":"cacf16f8d87a1df0ec2b7160f1917f444ba045c8ed39a1280d8748e49775dc82"} Oct 14 15:33:59 crc kubenswrapper[4945]: I1014 15:33:59.768821 4945 scope.go:117] "RemoveContainer" containerID="40d15dc11255c9246509b77850ad41e5fafbb3b18e210f748158b2b666b15891" Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.436952 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ce31-account-create-425x4" event={"ID":"82992e17-467c-46cf-8348-4248a6844b71","Type":"ContainerDied","Data":"defd92e1c8e62dbb2dd6aaab2b05e7a143bae7ddcd46ee9cd843d272408ec695"} Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.437239 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="defd92e1c8e62dbb2dd6aaab2b05e7a143bae7ddcd46ee9cd843d272408ec695" Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.439724 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7b358e-6730-43b7-bc2b-a154926a32de","Type":"ContainerStarted","Data":"287112c059673b0385b60fa2524cf356213529546faaf31e76018527916463ed"} Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.441081 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cdc7-account-create-qqwb7" event={"ID":"9dbf2952-a679-4420-910d-0ec3b3d6e1c1","Type":"ContainerDied","Data":"3e65ac0714910aa063f8088a7c689835f2521a4be7148d87eafd6d009956c077"} Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.441107 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e65ac0714910aa063f8088a7c689835f2521a4be7148d87eafd6d009956c077" Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.442428 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-4571-account-create-r2t9d" event={"ID":"d0c8cdd9-c134-4013-b110-388be1ead4db","Type":"ContainerDied","Data":"1b8e736c08adcc6d2cd420c6197ddf1391ba7adb1507b2ed40ecb7b3b9385a7e"} Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.442449 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b8e736c08adcc6d2cd420c6197ddf1391ba7adb1507b2ed40ecb7b3b9385a7e" Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.489547 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ce31-account-create-425x4" Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.500333 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cdc7-account-create-qqwb7" Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.504979 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-4571-account-create-r2t9d" Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.616938 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b2sd\" (UniqueName: \"kubernetes.io/projected/82992e17-467c-46cf-8348-4248a6844b71-kube-api-access-6b2sd\") pod \"82992e17-467c-46cf-8348-4248a6844b71\" (UID: \"82992e17-467c-46cf-8348-4248a6844b71\") " Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.617263 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xtjf\" (UniqueName: \"kubernetes.io/projected/9dbf2952-a679-4420-910d-0ec3b3d6e1c1-kube-api-access-2xtjf\") pod \"9dbf2952-a679-4420-910d-0ec3b3d6e1c1\" (UID: \"9dbf2952-a679-4420-910d-0ec3b3d6e1c1\") " Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.617522 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47jgw\" (UniqueName: \"kubernetes.io/projected/d0c8cdd9-c134-4013-b110-388be1ead4db-kube-api-access-47jgw\") pod \"d0c8cdd9-c134-4013-b110-388be1ead4db\" (UID: \"d0c8cdd9-c134-4013-b110-388be1ead4db\") " Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.620896 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dbf2952-a679-4420-910d-0ec3b3d6e1c1-kube-api-access-2xtjf" (OuterVolumeSpecName: "kube-api-access-2xtjf") pod "9dbf2952-a679-4420-910d-0ec3b3d6e1c1" (UID: "9dbf2952-a679-4420-910d-0ec3b3d6e1c1"). InnerVolumeSpecName "kube-api-access-2xtjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.621394 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82992e17-467c-46cf-8348-4248a6844b71-kube-api-access-6b2sd" (OuterVolumeSpecName: "kube-api-access-6b2sd") pod "82992e17-467c-46cf-8348-4248a6844b71" (UID: "82992e17-467c-46cf-8348-4248a6844b71"). InnerVolumeSpecName "kube-api-access-6b2sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.626990 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0c8cdd9-c134-4013-b110-388be1ead4db-kube-api-access-47jgw" (OuterVolumeSpecName: "kube-api-access-47jgw") pod "d0c8cdd9-c134-4013-b110-388be1ead4db" (UID: "d0c8cdd9-c134-4013-b110-388be1ead4db"). InnerVolumeSpecName "kube-api-access-47jgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.719408 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47jgw\" (UniqueName: \"kubernetes.io/projected/d0c8cdd9-c134-4013-b110-388be1ead4db-kube-api-access-47jgw\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.719440 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b2sd\" (UniqueName: \"kubernetes.io/projected/82992e17-467c-46cf-8348-4248a6844b71-kube-api-access-6b2sd\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:00 crc kubenswrapper[4945]: I1014 15:34:00.719449 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xtjf\" (UniqueName: \"kubernetes.io/projected/9dbf2952-a679-4420-910d-0ec3b3d6e1c1-kube-api-access-2xtjf\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:01 crc kubenswrapper[4945]: I1014 15:34:01.452461 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" event={"ID":"11506640-cea7-421c-878c-c4b449df880f","Type":"ContainerStarted","Data":"3603f131d46f535fe562976c5a3e4677fe3730d12b9ddc2cc971f48949614604"} Oct 14 15:34:01 crc kubenswrapper[4945]: I1014 15:34:01.453624 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:34:01 crc kubenswrapper[4945]: I1014 15:34:01.456004 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cdc7-account-create-qqwb7" Oct 14 15:34:01 crc kubenswrapper[4945]: I1014 15:34:01.456066 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-t252w" event={"ID":"048f7a68-56de-43d2-8ffb-33acacc4e2ec","Type":"ContainerStarted","Data":"4385009bc3e77fe426b5b28315418fe024d8bb647c7ee257cdcddbcbb423058d"} Oct 14 15:34:01 crc kubenswrapper[4945]: I1014 15:34:01.456520 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ce31-account-create-425x4" Oct 14 15:34:01 crc kubenswrapper[4945]: I1014 15:34:01.456850 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-4571-account-create-r2t9d" Oct 14 15:34:01 crc kubenswrapper[4945]: I1014 15:34:01.499381 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-inspector-db-sync-t252w" podStartSLOduration=2.867405384 podStartE2EDuration="6.499362143s" podCreationTimestamp="2025-10-14 15:33:55 +0000 UTC" firstStartedPulling="2025-10-14 15:33:56.745301603 +0000 UTC m=+1166.729349971" lastFinishedPulling="2025-10-14 15:34:00.377258342 +0000 UTC m=+1170.361306730" observedRunningTime="2025-10-14 15:34:01.493963369 +0000 UTC m=+1171.478011757" watchObservedRunningTime="2025-10-14 15:34:01.499362143 +0000 UTC m=+1171.483410511" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.220650 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-h5wlv"] Oct 14 15:34:02 crc kubenswrapper[4945]: E1014 15:34:02.221463 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0c8cdd9-c134-4013-b110-388be1ead4db" containerName="mariadb-account-create" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.221487 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0c8cdd9-c134-4013-b110-388be1ead4db" containerName="mariadb-account-create" Oct 14 15:34:02 crc kubenswrapper[4945]: E1014 15:34:02.221508 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82992e17-467c-46cf-8348-4248a6844b71" containerName="mariadb-account-create" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.221520 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="82992e17-467c-46cf-8348-4248a6844b71" containerName="mariadb-account-create" Oct 14 15:34:02 crc kubenswrapper[4945]: E1014 15:34:02.221561 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dbf2952-a679-4420-910d-0ec3b3d6e1c1" containerName="mariadb-account-create" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.221570 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dbf2952-a679-4420-910d-0ec3b3d6e1c1" containerName="mariadb-account-create" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.221786 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dbf2952-a679-4420-910d-0ec3b3d6e1c1" containerName="mariadb-account-create" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.221811 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="82992e17-467c-46cf-8348-4248a6844b71" containerName="mariadb-account-create" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.221836 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0c8cdd9-c134-4013-b110-388be1ead4db" containerName="mariadb-account-create" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.222591 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.223958 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.225302 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.227364 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wtwht" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.232045 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-h5wlv"] Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.356329 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-h5wlv\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.356687 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-scripts\") pod \"nova-cell0-conductor-db-sync-h5wlv\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.356817 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-config-data\") pod \"nova-cell0-conductor-db-sync-h5wlv\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.357011 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vxl6\" (UniqueName: \"kubernetes.io/projected/82939cb9-d9fc-4620-9a17-8990bd27bba8-kube-api-access-4vxl6\") pod \"nova-cell0-conductor-db-sync-h5wlv\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.465029 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-scripts\") pod \"nova-cell0-conductor-db-sync-h5wlv\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.465096 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-config-data\") pod \"nova-cell0-conductor-db-sync-h5wlv\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.465206 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vxl6\" (UniqueName: \"kubernetes.io/projected/82939cb9-d9fc-4620-9a17-8990bd27bba8-kube-api-access-4vxl6\") pod \"nova-cell0-conductor-db-sync-h5wlv\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.465236 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-h5wlv\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.471532 4945 generic.go:334] "Generic (PLEG): container finished" podID="048f7a68-56de-43d2-8ffb-33acacc4e2ec" containerID="4385009bc3e77fe426b5b28315418fe024d8bb647c7ee257cdcddbcbb423058d" exitCode=0 Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.471617 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-t252w" event={"ID":"048f7a68-56de-43d2-8ffb-33acacc4e2ec","Type":"ContainerDied","Data":"4385009bc3e77fe426b5b28315418fe024d8bb647c7ee257cdcddbcbb423058d"} Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.472613 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-h5wlv\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.482644 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-config-data\") pod \"nova-cell0-conductor-db-sync-h5wlv\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.486413 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-scripts\") pod \"nova-cell0-conductor-db-sync-h5wlv\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.490415 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vxl6\" (UniqueName: \"kubernetes.io/projected/82939cb9-d9fc-4620-9a17-8990bd27bba8-kube-api-access-4vxl6\") pod \"nova-cell0-conductor-db-sync-h5wlv\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.499773 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7b358e-6730-43b7-bc2b-a154926a32de","Type":"ContainerStarted","Data":"48f206cccb1b7bc26b1c884290635d1303d8e66212b36640ed5924b092c5b7a8"} Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.499815 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.540341 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:02 crc kubenswrapper[4945]: I1014 15:34:02.541746 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.568179183 podStartE2EDuration="6.541728306s" podCreationTimestamp="2025-10-14 15:33:56 +0000 UTC" firstStartedPulling="2025-10-14 15:33:57.65151966 +0000 UTC m=+1167.635568028" lastFinishedPulling="2025-10-14 15:34:01.625068773 +0000 UTC m=+1171.609117151" observedRunningTime="2025-10-14 15:34:02.537105274 +0000 UTC m=+1172.521153642" watchObservedRunningTime="2025-10-14 15:34:02.541728306 +0000 UTC m=+1172.525776664" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.067696 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-h5wlv"] Oct 14 15:34:03 crc kubenswrapper[4945]: W1014 15:34:03.074590 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82939cb9_d9fc_4620_9a17_8990bd27bba8.slice/crio-805b9e2f49529d3efe428f99dd8571f96b7ed030b97d797d4be71a17f43a2c9a WatchSource:0}: Error finding container 805b9e2f49529d3efe428f99dd8571f96b7ed030b97d797d4be71a17f43a2c9a: Status 404 returned error can't find the container with id 805b9e2f49529d3efe428f99dd8571f96b7ed030b97d797d4be71a17f43a2c9a Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.511025 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-h5wlv" event={"ID":"82939cb9-d9fc-4620-9a17-8990bd27bba8","Type":"ContainerStarted","Data":"805b9e2f49529d3efe428f99dd8571f96b7ed030b97d797d4be71a17f43a2c9a"} Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.885038 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.889014 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-scripts\") pod \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.889054 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/048f7a68-56de-43d2-8ffb-33acacc4e2ec-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.889129 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfssn\" (UniqueName: \"kubernetes.io/projected/048f7a68-56de-43d2-8ffb-33acacc4e2ec-kube-api-access-sfssn\") pod \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.889178 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/048f7a68-56de-43d2-8ffb-33acacc4e2ec-var-lib-ironic\") pod \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.889279 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-combined-ca-bundle\") pod \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.889325 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-config\") pod \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.889381 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/048f7a68-56de-43d2-8ffb-33acacc4e2ec-etc-podinfo\") pod \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\" (UID: \"048f7a68-56de-43d2-8ffb-33acacc4e2ec\") " Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.894554 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/048f7a68-56de-43d2-8ffb-33acacc4e2ec-var-lib-ironic" (OuterVolumeSpecName: "var-lib-ironic") pod "048f7a68-56de-43d2-8ffb-33acacc4e2ec" (UID: "048f7a68-56de-43d2-8ffb-33acacc4e2ec"). InnerVolumeSpecName "var-lib-ironic". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.895504 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-scripts" (OuterVolumeSpecName: "scripts") pod "048f7a68-56de-43d2-8ffb-33acacc4e2ec" (UID: "048f7a68-56de-43d2-8ffb-33acacc4e2ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.896142 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/048f7a68-56de-43d2-8ffb-33acacc4e2ec-var-lib-ironic-inspector-dhcp-hostsdir" (OuterVolumeSpecName: "var-lib-ironic-inspector-dhcp-hostsdir") pod "048f7a68-56de-43d2-8ffb-33acacc4e2ec" (UID: "048f7a68-56de-43d2-8ffb-33acacc4e2ec"). InnerVolumeSpecName "var-lib-ironic-inspector-dhcp-hostsdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.897446 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/048f7a68-56de-43d2-8ffb-33acacc4e2ec-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "048f7a68-56de-43d2-8ffb-33acacc4e2ec" (UID: "048f7a68-56de-43d2-8ffb-33acacc4e2ec"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.901084 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/048f7a68-56de-43d2-8ffb-33acacc4e2ec-kube-api-access-sfssn" (OuterVolumeSpecName: "kube-api-access-sfssn") pod "048f7a68-56de-43d2-8ffb-33acacc4e2ec" (UID: "048f7a68-56de-43d2-8ffb-33acacc4e2ec"). InnerVolumeSpecName "kube-api-access-sfssn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.941963 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-config" (OuterVolumeSpecName: "config") pod "048f7a68-56de-43d2-8ffb-33acacc4e2ec" (UID: "048f7a68-56de-43d2-8ffb-33acacc4e2ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.954588 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "048f7a68-56de-43d2-8ffb-33acacc4e2ec" (UID: "048f7a68-56de-43d2-8ffb-33acacc4e2ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.992912 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.992944 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.992973 4945 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/048f7a68-56de-43d2-8ffb-33acacc4e2ec-etc-podinfo\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.992982 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/048f7a68-56de-43d2-8ffb-33acacc4e2ec-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.992991 4945 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/048f7a68-56de-43d2-8ffb-33acacc4e2ec-var-lib-ironic-inspector-dhcp-hostsdir\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.993001 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfssn\" (UniqueName: \"kubernetes.io/projected/048f7a68-56de-43d2-8ffb-33acacc4e2ec-kube-api-access-sfssn\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:03 crc kubenswrapper[4945]: I1014 15:34:03.993011 4945 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/048f7a68-56de-43d2-8ffb-33acacc4e2ec-var-lib-ironic\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:04 crc kubenswrapper[4945]: I1014 15:34:04.523474 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-t252w" event={"ID":"048f7a68-56de-43d2-8ffb-33acacc4e2ec","Type":"ContainerDied","Data":"cf81a0f5fb2f18581f3579f5213835ab144d779285b57e3e090d146b993cdcd9"} Oct 14 15:34:04 crc kubenswrapper[4945]: I1014 15:34:04.523819 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf81a0f5fb2f18581f3579f5213835ab144d779285b57e3e090d146b993cdcd9" Oct 14 15:34:04 crc kubenswrapper[4945]: I1014 15:34:04.523506 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-t252w" Oct 14 15:34:04 crc kubenswrapper[4945]: I1014 15:34:04.526904 4945 generic.go:334] "Generic (PLEG): container finished" podID="11506640-cea7-421c-878c-c4b449df880f" containerID="3603f131d46f535fe562976c5a3e4677fe3730d12b9ddc2cc971f48949614604" exitCode=1 Oct 14 15:34:04 crc kubenswrapper[4945]: I1014 15:34:04.526949 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" event={"ID":"11506640-cea7-421c-878c-c4b449df880f","Type":"ContainerDied","Data":"3603f131d46f535fe562976c5a3e4677fe3730d12b9ddc2cc971f48949614604"} Oct 14 15:34:04 crc kubenswrapper[4945]: I1014 15:34:04.526981 4945 scope.go:117] "RemoveContainer" containerID="40d15dc11255c9246509b77850ad41e5fafbb3b18e210f748158b2b666b15891" Oct 14 15:34:04 crc kubenswrapper[4945]: I1014 15:34:04.527766 4945 scope.go:117] "RemoveContainer" containerID="3603f131d46f535fe562976c5a3e4677fe3730d12b9ddc2cc971f48949614604" Oct 14 15:34:04 crc kubenswrapper[4945]: E1014 15:34:04.529483 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-648bd876cc-srkfr_openstack(11506640-cea7-421c-878c-c4b449df880f)\"" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" podUID="11506640-cea7-421c-878c-c4b449df880f" Oct 14 15:34:05 crc kubenswrapper[4945]: I1014 15:34:05.042789 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:34:05 crc kubenswrapper[4945]: I1014 15:34:05.043193 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="ceilometer-central-agent" containerID="cri-o://5f5ad7834f6e37bd63f4a76a3714f09174f167cdb441d1374f580b32694f2771" gracePeriod=30 Oct 14 15:34:05 crc kubenswrapper[4945]: I1014 15:34:05.043271 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="proxy-httpd" containerID="cri-o://48f206cccb1b7bc26b1c884290635d1303d8e66212b36640ed5924b092c5b7a8" gracePeriod=30 Oct 14 15:34:05 crc kubenswrapper[4945]: I1014 15:34:05.043323 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="sg-core" containerID="cri-o://287112c059673b0385b60fa2524cf356213529546faaf31e76018527916463ed" gracePeriod=30 Oct 14 15:34:05 crc kubenswrapper[4945]: I1014 15:34:05.043340 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="ceilometer-notification-agent" containerID="cri-o://cacf16f8d87a1df0ec2b7160f1917f444ba045c8ed39a1280d8748e49775dc82" gracePeriod=30 Oct 14 15:34:05 crc kubenswrapper[4945]: I1014 15:34:05.539314 4945 generic.go:334] "Generic (PLEG): container finished" podID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerID="48f206cccb1b7bc26b1c884290635d1303d8e66212b36640ed5924b092c5b7a8" exitCode=0 Oct 14 15:34:05 crc kubenswrapper[4945]: I1014 15:34:05.539381 4945 generic.go:334] "Generic (PLEG): container finished" podID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerID="287112c059673b0385b60fa2524cf356213529546faaf31e76018527916463ed" exitCode=2 Oct 14 15:34:05 crc kubenswrapper[4945]: I1014 15:34:05.539389 4945 generic.go:334] "Generic (PLEG): container finished" podID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerID="cacf16f8d87a1df0ec2b7160f1917f444ba045c8ed39a1280d8748e49775dc82" exitCode=0 Oct 14 15:34:05 crc kubenswrapper[4945]: I1014 15:34:05.539421 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7b358e-6730-43b7-bc2b-a154926a32de","Type":"ContainerDied","Data":"48f206cccb1b7bc26b1c884290635d1303d8e66212b36640ed5924b092c5b7a8"} Oct 14 15:34:05 crc kubenswrapper[4945]: I1014 15:34:05.539446 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7b358e-6730-43b7-bc2b-a154926a32de","Type":"ContainerDied","Data":"287112c059673b0385b60fa2524cf356213529546faaf31e76018527916463ed"} Oct 14 15:34:05 crc kubenswrapper[4945]: I1014 15:34:05.539461 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7b358e-6730-43b7-bc2b-a154926a32de","Type":"ContainerDied","Data":"cacf16f8d87a1df0ec2b7160f1917f444ba045c8ed39a1280d8748e49775dc82"} Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.148157 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-0"] Oct 14 15:34:06 crc kubenswrapper[4945]: E1014 15:34:06.148936 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048f7a68-56de-43d2-8ffb-33acacc4e2ec" containerName="ironic-inspector-db-sync" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.148959 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="048f7a68-56de-43d2-8ffb-33acacc4e2ec" containerName="ironic-inspector-db-sync" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.149255 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="048f7a68-56de-43d2-8ffb-33acacc4e2ec" containerName="ironic-inspector-db-sync" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.182128 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.187265 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.187751 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.213979 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.270015 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.271930 4945 scope.go:117] "RemoveContainer" containerID="3603f131d46f535fe562976c5a3e4677fe3730d12b9ddc2cc971f48949614604" Oct 14 15:34:06 crc kubenswrapper[4945]: E1014 15:34:06.272224 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-648bd876cc-srkfr_openstack(11506640-cea7-421c-878c-c4b449df880f)\"" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" podUID="11506640-cea7-421c-878c-c4b449df880f" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.333852 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-config\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.334191 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/c097cadc-aa5c-43c1-a2b1-1af6739d4086-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.334216 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-scripts\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.334241 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfmhv\" (UniqueName: \"kubernetes.io/projected/c097cadc-aa5c-43c1-a2b1-1af6739d4086-kube-api-access-vfmhv\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.334267 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/c097cadc-aa5c-43c1-a2b1-1af6739d4086-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.334289 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.334303 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/c097cadc-aa5c-43c1-a2b1-1af6739d4086-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.435743 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/c097cadc-aa5c-43c1-a2b1-1af6739d4086-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.435794 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-scripts\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.435834 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfmhv\" (UniqueName: \"kubernetes.io/projected/c097cadc-aa5c-43c1-a2b1-1af6739d4086-kube-api-access-vfmhv\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.435860 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/c097cadc-aa5c-43c1-a2b1-1af6739d4086-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.435899 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.435916 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/c097cadc-aa5c-43c1-a2b1-1af6739d4086-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.436013 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-config\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.436732 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/c097cadc-aa5c-43c1-a2b1-1af6739d4086-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.436957 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/c097cadc-aa5c-43c1-a2b1-1af6739d4086-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.445542 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-scripts\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.447187 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/c097cadc-aa5c-43c1-a2b1-1af6739d4086-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.447319 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.448953 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-config\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.454746 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfmhv\" (UniqueName: \"kubernetes.io/projected/c097cadc-aa5c-43c1-a2b1-1af6739d4086-kube-api-access-vfmhv\") pod \"ironic-inspector-0\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:06 crc kubenswrapper[4945]: I1014 15:34:06.560464 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 14 15:34:08 crc kubenswrapper[4945]: I1014 15:34:08.588349 4945 generic.go:334] "Generic (PLEG): container finished" podID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerID="5f5ad7834f6e37bd63f4a76a3714f09174f167cdb441d1374f580b32694f2771" exitCode=0 Oct 14 15:34:08 crc kubenswrapper[4945]: I1014 15:34:08.588416 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7b358e-6730-43b7-bc2b-a154926a32de","Type":"ContainerDied","Data":"5f5ad7834f6e37bd63f4a76a3714f09174f167cdb441d1374f580b32694f2771"} Oct 14 15:34:09 crc kubenswrapper[4945]: I1014 15:34:09.160451 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-0"] Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.809665 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.943444 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-sg-core-conf-yaml\") pod \"2a7b358e-6730-43b7-bc2b-a154926a32de\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.943548 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7b358e-6730-43b7-bc2b-a154926a32de-log-httpd\") pod \"2a7b358e-6730-43b7-bc2b-a154926a32de\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.943609 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-config-data\") pod \"2a7b358e-6730-43b7-bc2b-a154926a32de\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.943668 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7b358e-6730-43b7-bc2b-a154926a32de-run-httpd\") pod \"2a7b358e-6730-43b7-bc2b-a154926a32de\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.943702 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74phm\" (UniqueName: \"kubernetes.io/projected/2a7b358e-6730-43b7-bc2b-a154926a32de-kube-api-access-74phm\") pod \"2a7b358e-6730-43b7-bc2b-a154926a32de\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.943731 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-scripts\") pod \"2a7b358e-6730-43b7-bc2b-a154926a32de\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.943790 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-combined-ca-bundle\") pod \"2a7b358e-6730-43b7-bc2b-a154926a32de\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.944179 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a7b358e-6730-43b7-bc2b-a154926a32de-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2a7b358e-6730-43b7-bc2b-a154926a32de" (UID: "2a7b358e-6730-43b7-bc2b-a154926a32de"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.944307 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a7b358e-6730-43b7-bc2b-a154926a32de-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2a7b358e-6730-43b7-bc2b-a154926a32de" (UID: "2a7b358e-6730-43b7-bc2b-a154926a32de"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.944323 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7b358e-6730-43b7-bc2b-a154926a32de-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.948113 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a7b358e-6730-43b7-bc2b-a154926a32de-kube-api-access-74phm" (OuterVolumeSpecName: "kube-api-access-74phm") pod "2a7b358e-6730-43b7-bc2b-a154926a32de" (UID: "2a7b358e-6730-43b7-bc2b-a154926a32de"). InnerVolumeSpecName "kube-api-access-74phm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.950188 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-scripts" (OuterVolumeSpecName: "scripts") pod "2a7b358e-6730-43b7-bc2b-a154926a32de" (UID: "2a7b358e-6730-43b7-bc2b-a154926a32de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.977013 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2a7b358e-6730-43b7-bc2b-a154926a32de" (UID: "2a7b358e-6730-43b7-bc2b-a154926a32de"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:11 crc kubenswrapper[4945]: I1014 15:34:11.998540 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-0"] Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.024155 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a7b358e-6730-43b7-bc2b-a154926a32de" (UID: "2a7b358e-6730-43b7-bc2b-a154926a32de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.045152 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-config-data" (OuterVolumeSpecName: "config-data") pod "2a7b358e-6730-43b7-bc2b-a154926a32de" (UID: "2a7b358e-6730-43b7-bc2b-a154926a32de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.045867 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-config-data\") pod \"2a7b358e-6730-43b7-bc2b-a154926a32de\" (UID: \"2a7b358e-6730-43b7-bc2b-a154926a32de\") " Oct 14 15:34:12 crc kubenswrapper[4945]: W1014 15:34:12.045956 4945 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/2a7b358e-6730-43b7-bc2b-a154926a32de/volumes/kubernetes.io~secret/config-data Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.045968 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-config-data" (OuterVolumeSpecName: "config-data") pod "2a7b358e-6730-43b7-bc2b-a154926a32de" (UID: "2a7b358e-6730-43b7-bc2b-a154926a32de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.046377 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.046397 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.046411 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7b358e-6730-43b7-bc2b-a154926a32de-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.046422 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.046433 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74phm\" (UniqueName: \"kubernetes.io/projected/2a7b358e-6730-43b7-bc2b-a154926a32de-kube-api-access-74phm\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.046447 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a7b358e-6730-43b7-bc2b-a154926a32de-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.632191 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7b358e-6730-43b7-bc2b-a154926a32de","Type":"ContainerDied","Data":"693c94bd3c6022c1b90f23674544d526844c0fbded319528569ee9a68df0eebc"} Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.632485 4945 scope.go:117] "RemoveContainer" containerID="48f206cccb1b7bc26b1c884290635d1303d8e66212b36640ed5924b092c5b7a8" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.632391 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.634593 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"c097cadc-aa5c-43c1-a2b1-1af6739d4086","Type":"ContainerStarted","Data":"307dba8a52dc1fdeaff859cd54922ea3997bd646f0e256c3ebb53cc7703e2fdf"} Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.636700 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-h5wlv" event={"ID":"82939cb9-d9fc-4620-9a17-8990bd27bba8","Type":"ContainerStarted","Data":"3b038afb3f1c6faadf96de859d0a6d8f9c7dca8a10fee57f405143963947a5a6"} Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.655667 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-h5wlv" podStartSLOduration=2.209890379 podStartE2EDuration="10.655648951s" podCreationTimestamp="2025-10-14 15:34:02 +0000 UTC" firstStartedPulling="2025-10-14 15:34:03.077097913 +0000 UTC m=+1173.061146281" lastFinishedPulling="2025-10-14 15:34:11.522856485 +0000 UTC m=+1181.506904853" observedRunningTime="2025-10-14 15:34:12.64862854 +0000 UTC m=+1182.632676908" watchObservedRunningTime="2025-10-14 15:34:12.655648951 +0000 UTC m=+1182.639697319" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.662828 4945 scope.go:117] "RemoveContainer" containerID="287112c059673b0385b60fa2524cf356213529546faaf31e76018527916463ed" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.681076 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.701274 4945 scope.go:117] "RemoveContainer" containerID="cacf16f8d87a1df0ec2b7160f1917f444ba045c8ed39a1280d8748e49775dc82" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.707492 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.715245 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:34:12 crc kubenswrapper[4945]: E1014 15:34:12.715665 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="sg-core" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.715678 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="sg-core" Oct 14 15:34:12 crc kubenswrapper[4945]: E1014 15:34:12.715694 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="ceilometer-notification-agent" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.715700 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="ceilometer-notification-agent" Oct 14 15:34:12 crc kubenswrapper[4945]: E1014 15:34:12.715709 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="proxy-httpd" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.715715 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="proxy-httpd" Oct 14 15:34:12 crc kubenswrapper[4945]: E1014 15:34:12.715723 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="ceilometer-central-agent" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.715729 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="ceilometer-central-agent" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.715985 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="ceilometer-notification-agent" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.715995 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="sg-core" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.716009 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="ceilometer-central-agent" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.716015 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" containerName="proxy-httpd" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.727318 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.727561 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.729929 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.731106 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.765345 4945 scope.go:117] "RemoveContainer" containerID="5f5ad7834f6e37bd63f4a76a3714f09174f167cdb441d1374f580b32694f2771" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.790656 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a7b358e-6730-43b7-bc2b-a154926a32de" path="/var/lib/kubelet/pods/2a7b358e-6730-43b7-bc2b-a154926a32de/volumes" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.863812 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-run-httpd\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.864188 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kh4x\" (UniqueName: \"kubernetes.io/projected/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-kube-api-access-2kh4x\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.864216 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-log-httpd\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.864787 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.864934 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-config-data\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.865771 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-scripts\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.865800 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.967380 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-scripts\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.967463 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.967489 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-run-httpd\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.968220 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-run-httpd\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.968509 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kh4x\" (UniqueName: \"kubernetes.io/projected/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-kube-api-access-2kh4x\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.968579 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-log-httpd\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.968736 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.968780 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-config-data\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.969196 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-log-httpd\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.974700 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-scripts\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.974716 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-config-data\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.977976 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.985809 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:12 crc kubenswrapper[4945]: I1014 15:34:12.986267 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kh4x\" (UniqueName: \"kubernetes.io/projected/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-kube-api-access-2kh4x\") pod \"ceilometer-0\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " pod="openstack/ceilometer-0" Oct 14 15:34:13 crc kubenswrapper[4945]: I1014 15:34:13.058070 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:34:13 crc kubenswrapper[4945]: I1014 15:34:13.489430 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:34:13 crc kubenswrapper[4945]: I1014 15:34:13.649683 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd04b8b0-f8ab-4851-8303-c9ab3286bf58","Type":"ContainerStarted","Data":"81ec76b7ddd52aa266874d375ecb8268e9807fbbfb6c0cebca98f551d01365ca"} Oct 14 15:34:14 crc kubenswrapper[4945]: I1014 15:34:14.667686 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd04b8b0-f8ab-4851-8303-c9ab3286bf58","Type":"ContainerStarted","Data":"efbbf41db8b3ea1be2d941267bb61587b08fe52e37bf4ff26d840f59dc1cedad"} Oct 14 15:34:16 crc kubenswrapper[4945]: I1014 15:34:16.688531 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd04b8b0-f8ab-4851-8303-c9ab3286bf58","Type":"ContainerStarted","Data":"766f810296ec7a5a93ef1b347b3515dc99539fc1028d898e008f6876c2c041ca"} Oct 14 15:34:17 crc kubenswrapper[4945]: I1014 15:34:17.709235 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd04b8b0-f8ab-4851-8303-c9ab3286bf58","Type":"ContainerStarted","Data":"4da9d23841a555fd547a272e8fc433ccbaa0dd14c62e71ff5b0495896e9af169"} Oct 14 15:34:18 crc kubenswrapper[4945]: I1014 15:34:18.762845 4945 scope.go:117] "RemoveContainer" containerID="3603f131d46f535fe562976c5a3e4677fe3730d12b9ddc2cc971f48949614604" Oct 14 15:34:18 crc kubenswrapper[4945]: E1014 15:34:18.763375 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-648bd876cc-srkfr_openstack(11506640-cea7-421c-878c-c4b449df880f)\"" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" podUID="11506640-cea7-421c-878c-c4b449df880f" Oct 14 15:34:19 crc kubenswrapper[4945]: I1014 15:34:19.732607 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd04b8b0-f8ab-4851-8303-c9ab3286bf58","Type":"ContainerStarted","Data":"e2fc55e977d5b6625355202d173ead6e7f1899f58d98552f0ccc6b6f92e1bea9"} Oct 14 15:34:19 crc kubenswrapper[4945]: I1014 15:34:19.732808 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 15:34:19 crc kubenswrapper[4945]: I1014 15:34:19.768863 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.772591227 podStartE2EDuration="7.76882508s" podCreationTimestamp="2025-10-14 15:34:12 +0000 UTC" firstStartedPulling="2025-10-14 15:34:13.498134766 +0000 UTC m=+1183.482183134" lastFinishedPulling="2025-10-14 15:34:18.494368609 +0000 UTC m=+1188.478416987" observedRunningTime="2025-10-14 15:34:19.759559156 +0000 UTC m=+1189.743607574" watchObservedRunningTime="2025-10-14 15:34:19.76882508 +0000 UTC m=+1189.752873458" Oct 14 15:34:19 crc kubenswrapper[4945]: I1014 15:34:19.936422 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:34:21 crc kubenswrapper[4945]: I1014 15:34:21.755707 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="ceilometer-central-agent" containerID="cri-o://efbbf41db8b3ea1be2d941267bb61587b08fe52e37bf4ff26d840f59dc1cedad" gracePeriod=30 Oct 14 15:34:21 crc kubenswrapper[4945]: I1014 15:34:21.755741 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="sg-core" containerID="cri-o://4da9d23841a555fd547a272e8fc433ccbaa0dd14c62e71ff5b0495896e9af169" gracePeriod=30 Oct 14 15:34:21 crc kubenswrapper[4945]: I1014 15:34:21.755723 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="proxy-httpd" containerID="cri-o://e2fc55e977d5b6625355202d173ead6e7f1899f58d98552f0ccc6b6f92e1bea9" gracePeriod=30 Oct 14 15:34:21 crc kubenswrapper[4945]: I1014 15:34:21.755800 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="ceilometer-notification-agent" containerID="cri-o://766f810296ec7a5a93ef1b347b3515dc99539fc1028d898e008f6876c2c041ca" gracePeriod=30 Oct 14 15:34:22 crc kubenswrapper[4945]: I1014 15:34:22.765682 4945 generic.go:334] "Generic (PLEG): container finished" podID="82939cb9-d9fc-4620-9a17-8990bd27bba8" containerID="3b038afb3f1c6faadf96de859d0a6d8f9c7dca8a10fee57f405143963947a5a6" exitCode=0 Oct 14 15:34:22 crc kubenswrapper[4945]: I1014 15:34:22.770305 4945 generic.go:334] "Generic (PLEG): container finished" podID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerID="e2fc55e977d5b6625355202d173ead6e7f1899f58d98552f0ccc6b6f92e1bea9" exitCode=0 Oct 14 15:34:22 crc kubenswrapper[4945]: I1014 15:34:22.770345 4945 generic.go:334] "Generic (PLEG): container finished" podID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerID="4da9d23841a555fd547a272e8fc433ccbaa0dd14c62e71ff5b0495896e9af169" exitCode=2 Oct 14 15:34:22 crc kubenswrapper[4945]: I1014 15:34:22.772643 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-h5wlv" event={"ID":"82939cb9-d9fc-4620-9a17-8990bd27bba8","Type":"ContainerDied","Data":"3b038afb3f1c6faadf96de859d0a6d8f9c7dca8a10fee57f405143963947a5a6"} Oct 14 15:34:22 crc kubenswrapper[4945]: I1014 15:34:22.772690 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd04b8b0-f8ab-4851-8303-c9ab3286bf58","Type":"ContainerDied","Data":"e2fc55e977d5b6625355202d173ead6e7f1899f58d98552f0ccc6b6f92e1bea9"} Oct 14 15:34:22 crc kubenswrapper[4945]: I1014 15:34:22.772703 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd04b8b0-f8ab-4851-8303-c9ab3286bf58","Type":"ContainerDied","Data":"4da9d23841a555fd547a272e8fc433ccbaa0dd14c62e71ff5b0495896e9af169"} Oct 14 15:34:23 crc kubenswrapper[4945]: I1014 15:34:23.780691 4945 generic.go:334] "Generic (PLEG): container finished" podID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerID="766f810296ec7a5a93ef1b347b3515dc99539fc1028d898e008f6876c2c041ca" exitCode=0 Oct 14 15:34:23 crc kubenswrapper[4945]: I1014 15:34:23.780719 4945 generic.go:334] "Generic (PLEG): container finished" podID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerID="efbbf41db8b3ea1be2d941267bb61587b08fe52e37bf4ff26d840f59dc1cedad" exitCode=0 Oct 14 15:34:23 crc kubenswrapper[4945]: I1014 15:34:23.780753 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd04b8b0-f8ab-4851-8303-c9ab3286bf58","Type":"ContainerDied","Data":"766f810296ec7a5a93ef1b347b3515dc99539fc1028d898e008f6876c2c041ca"} Oct 14 15:34:23 crc kubenswrapper[4945]: I1014 15:34:23.780784 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd04b8b0-f8ab-4851-8303-c9ab3286bf58","Type":"ContainerDied","Data":"efbbf41db8b3ea1be2d941267bb61587b08fe52e37bf4ff26d840f59dc1cedad"} Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.115455 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.209791 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-scripts\") pod \"82939cb9-d9fc-4620-9a17-8990bd27bba8\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.209928 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vxl6\" (UniqueName: \"kubernetes.io/projected/82939cb9-d9fc-4620-9a17-8990bd27bba8-kube-api-access-4vxl6\") pod \"82939cb9-d9fc-4620-9a17-8990bd27bba8\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.209990 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-combined-ca-bundle\") pod \"82939cb9-d9fc-4620-9a17-8990bd27bba8\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.210057 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-config-data\") pod \"82939cb9-d9fc-4620-9a17-8990bd27bba8\" (UID: \"82939cb9-d9fc-4620-9a17-8990bd27bba8\") " Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.216407 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82939cb9-d9fc-4620-9a17-8990bd27bba8-kube-api-access-4vxl6" (OuterVolumeSpecName: "kube-api-access-4vxl6") pod "82939cb9-d9fc-4620-9a17-8990bd27bba8" (UID: "82939cb9-d9fc-4620-9a17-8990bd27bba8"). InnerVolumeSpecName "kube-api-access-4vxl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.216491 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-scripts" (OuterVolumeSpecName: "scripts") pod "82939cb9-d9fc-4620-9a17-8990bd27bba8" (UID: "82939cb9-d9fc-4620-9a17-8990bd27bba8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.249818 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82939cb9-d9fc-4620-9a17-8990bd27bba8" (UID: "82939cb9-d9fc-4620-9a17-8990bd27bba8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.256056 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-config-data" (OuterVolumeSpecName: "config-data") pod "82939cb9-d9fc-4620-9a17-8990bd27bba8" (UID: "82939cb9-d9fc-4620-9a17-8990bd27bba8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.311507 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.311541 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.311551 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vxl6\" (UniqueName: \"kubernetes.io/projected/82939cb9-d9fc-4620-9a17-8990bd27bba8-kube-api-access-4vxl6\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.311560 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82939cb9-d9fc-4620-9a17-8990bd27bba8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.805615 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-h5wlv" event={"ID":"82939cb9-d9fc-4620-9a17-8990bd27bba8","Type":"ContainerDied","Data":"805b9e2f49529d3efe428f99dd8571f96b7ed030b97d797d4be71a17f43a2c9a"} Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.805670 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="805b9e2f49529d3efe428f99dd8571f96b7ed030b97d797d4be71a17f43a2c9a" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.805687 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-h5wlv" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.915283 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 15:34:24 crc kubenswrapper[4945]: E1014 15:34:24.915913 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82939cb9-d9fc-4620-9a17-8990bd27bba8" containerName="nova-cell0-conductor-db-sync" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.915937 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="82939cb9-d9fc-4620-9a17-8990bd27bba8" containerName="nova-cell0-conductor-db-sync" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.916280 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="82939cb9-d9fc-4620-9a17-8990bd27bba8" containerName="nova-cell0-conductor-db-sync" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.917122 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.919338 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wtwht" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.919485 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 15:34:24 crc kubenswrapper[4945]: I1014 15:34:24.935441 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 15:34:25 crc kubenswrapper[4945]: I1014 15:34:25.028760 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdn72\" (UniqueName: \"kubernetes.io/projected/f4bf8e7b-6b0a-48da-831c-a9c307af75b9-kube-api-access-pdn72\") pod \"nova-cell0-conductor-0\" (UID: \"f4bf8e7b-6b0a-48da-831c-a9c307af75b9\") " pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:25 crc kubenswrapper[4945]: I1014 15:34:25.028872 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4bf8e7b-6b0a-48da-831c-a9c307af75b9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f4bf8e7b-6b0a-48da-831c-a9c307af75b9\") " pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:25 crc kubenswrapper[4945]: I1014 15:34:25.029178 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4bf8e7b-6b0a-48da-831c-a9c307af75b9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f4bf8e7b-6b0a-48da-831c-a9c307af75b9\") " pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:25 crc kubenswrapper[4945]: I1014 15:34:25.131176 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdn72\" (UniqueName: \"kubernetes.io/projected/f4bf8e7b-6b0a-48da-831c-a9c307af75b9-kube-api-access-pdn72\") pod \"nova-cell0-conductor-0\" (UID: \"f4bf8e7b-6b0a-48da-831c-a9c307af75b9\") " pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:25 crc kubenswrapper[4945]: I1014 15:34:25.131258 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4bf8e7b-6b0a-48da-831c-a9c307af75b9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f4bf8e7b-6b0a-48da-831c-a9c307af75b9\") " pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:25 crc kubenswrapper[4945]: I1014 15:34:25.131457 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4bf8e7b-6b0a-48da-831c-a9c307af75b9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f4bf8e7b-6b0a-48da-831c-a9c307af75b9\") " pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:25 crc kubenswrapper[4945]: I1014 15:34:25.139686 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4bf8e7b-6b0a-48da-831c-a9c307af75b9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f4bf8e7b-6b0a-48da-831c-a9c307af75b9\") " pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:25 crc kubenswrapper[4945]: I1014 15:34:25.141519 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4bf8e7b-6b0a-48da-831c-a9c307af75b9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f4bf8e7b-6b0a-48da-831c-a9c307af75b9\") " pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:25 crc kubenswrapper[4945]: I1014 15:34:25.148405 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdn72\" (UniqueName: \"kubernetes.io/projected/f4bf8e7b-6b0a-48da-831c-a9c307af75b9-kube-api-access-pdn72\") pod \"nova-cell0-conductor-0\" (UID: \"f4bf8e7b-6b0a-48da-831c-a9c307af75b9\") " pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:25 crc kubenswrapper[4945]: I1014 15:34:25.236901 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:30 crc kubenswrapper[4945]: I1014 15:34:30.869723 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd04b8b0-f8ab-4851-8303-c9ab3286bf58","Type":"ContainerDied","Data":"81ec76b7ddd52aa266874d375ecb8268e9807fbbfb6c0cebca98f551d01365ca"} Oct 14 15:34:30 crc kubenswrapper[4945]: I1014 15:34:30.870760 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81ec76b7ddd52aa266874d375ecb8268e9807fbbfb6c0cebca98f551d01365ca" Oct 14 15:34:30 crc kubenswrapper[4945]: I1014 15:34:30.875111 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.012263 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-sg-core-conf-yaml\") pod \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.012399 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-run-httpd\") pod \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.012461 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-scripts\") pod \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.012492 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kh4x\" (UniqueName: \"kubernetes.io/projected/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-kube-api-access-2kh4x\") pod \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.012539 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-combined-ca-bundle\") pod \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.012652 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-log-httpd\") pod \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.012680 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-config-data\") pod \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\" (UID: \"fd04b8b0-f8ab-4851-8303-c9ab3286bf58\") " Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.016055 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fd04b8b0-f8ab-4851-8303-c9ab3286bf58" (UID: "fd04b8b0-f8ab-4851-8303-c9ab3286bf58"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.016637 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fd04b8b0-f8ab-4851-8303-c9ab3286bf58" (UID: "fd04b8b0-f8ab-4851-8303-c9ab3286bf58"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.021912 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-kube-api-access-2kh4x" (OuterVolumeSpecName: "kube-api-access-2kh4x") pod "fd04b8b0-f8ab-4851-8303-c9ab3286bf58" (UID: "fd04b8b0-f8ab-4851-8303-c9ab3286bf58"). InnerVolumeSpecName "kube-api-access-2kh4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.023294 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-scripts" (OuterVolumeSpecName: "scripts") pod "fd04b8b0-f8ab-4851-8303-c9ab3286bf58" (UID: "fd04b8b0-f8ab-4851-8303-c9ab3286bf58"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.065649 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fd04b8b0-f8ab-4851-8303-c9ab3286bf58" (UID: "fd04b8b0-f8ab-4851-8303-c9ab3286bf58"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.104930 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd04b8b0-f8ab-4851-8303-c9ab3286bf58" (UID: "fd04b8b0-f8ab-4851-8303-c9ab3286bf58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.114836 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kh4x\" (UniqueName: \"kubernetes.io/projected/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-kube-api-access-2kh4x\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.114882 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.114900 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.114911 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.114923 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.114934 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.149023 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-config-data" (OuterVolumeSpecName: "config-data") pod "fd04b8b0-f8ab-4851-8303-c9ab3286bf58" (UID: "fd04b8b0-f8ab-4851-8303-c9ab3286bf58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.216811 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd04b8b0-f8ab-4851-8303-c9ab3286bf58-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.297582 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.880853 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f4bf8e7b-6b0a-48da-831c-a9c307af75b9","Type":"ContainerStarted","Data":"b8a16990d79b60c068f1fee338a6533346b65d618d3a4b26503aad8c09829d80"} Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.881484 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.881959 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f4bf8e7b-6b0a-48da-831c-a9c307af75b9","Type":"ContainerStarted","Data":"e9fa07019aaf7f5f4b3e61acc2291a128b0ad97da9847eca08fc4d56603de814"} Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.884595 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"042f4fc3-10ec-450e-b3cf-77a11d71a8b7","Type":"ContainerStarted","Data":"62ec4fa672b977907ddecbe53d8a70d1aa572a93de1d45c57b587aea8d8f33c6"} Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.887184 4945 generic.go:334] "Generic (PLEG): container finished" podID="c097cadc-aa5c-43c1-a2b1-1af6739d4086" containerID="bd72f628ad6f4818445a894bd5041c2c37945429b577069463c2ab0af9674d95" exitCode=0 Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.887293 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.896220 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"c097cadc-aa5c-43c1-a2b1-1af6739d4086","Type":"ContainerDied","Data":"bd72f628ad6f4818445a894bd5041c2c37945429b577069463c2ab0af9674d95"} Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.904423 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=7.90440468 podStartE2EDuration="7.90440468s" podCreationTimestamp="2025-10-14 15:34:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:34:31.895096774 +0000 UTC m=+1201.879145152" watchObservedRunningTime="2025-10-14 15:34:31.90440468 +0000 UTC m=+1201.888453048" Oct 14 15:34:31 crc kubenswrapper[4945]: I1014 15:34:31.996855 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.027085 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.043566 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:34:32 crc kubenswrapper[4945]: E1014 15:34:32.043974 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="sg-core" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.043998 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="sg-core" Oct 14 15:34:32 crc kubenswrapper[4945]: E1014 15:34:32.044018 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="ceilometer-central-agent" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.044023 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="ceilometer-central-agent" Oct 14 15:34:32 crc kubenswrapper[4945]: E1014 15:34:32.044127 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="proxy-httpd" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.044137 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="proxy-httpd" Oct 14 15:34:32 crc kubenswrapper[4945]: E1014 15:34:32.044147 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="ceilometer-notification-agent" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.044153 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="ceilometer-notification-agent" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.044323 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="proxy-httpd" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.044343 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="sg-core" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.044356 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="ceilometer-central-agent" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.044376 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" containerName="ceilometer-notification-agent" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.046105 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.048145 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.048350 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.049462 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.138377 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.138428 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp262\" (UniqueName: \"kubernetes.io/projected/9fdab3dc-df19-4f56-a226-8afb55579e2e-kube-api-access-xp262\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.138480 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-scripts\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.138508 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.138579 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fdab3dc-df19-4f56-a226-8afb55579e2e-run-httpd\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.138665 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fdab3dc-df19-4f56-a226-8afb55579e2e-log-httpd\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.138815 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-config-data\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.241117 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp262\" (UniqueName: \"kubernetes.io/projected/9fdab3dc-df19-4f56-a226-8afb55579e2e-kube-api-access-xp262\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.241206 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-scripts\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.241243 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.241265 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fdab3dc-df19-4f56-a226-8afb55579e2e-run-httpd\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.241300 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fdab3dc-df19-4f56-a226-8afb55579e2e-log-httpd\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.241355 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-config-data\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.241457 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.243315 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fdab3dc-df19-4f56-a226-8afb55579e2e-log-httpd\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.243930 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fdab3dc-df19-4f56-a226-8afb55579e2e-run-httpd\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.246652 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.246732 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.247341 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-scripts\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.258651 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-config-data\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.264336 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp262\" (UniqueName: \"kubernetes.io/projected/9fdab3dc-df19-4f56-a226-8afb55579e2e-kube-api-access-xp262\") pod \"ceilometer-0\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.323713 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.365633 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.444352 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfmhv\" (UniqueName: \"kubernetes.io/projected/c097cadc-aa5c-43c1-a2b1-1af6739d4086-kube-api-access-vfmhv\") pod \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.444473 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/c097cadc-aa5c-43c1-a2b1-1af6739d4086-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.444520 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/c097cadc-aa5c-43c1-a2b1-1af6739d4086-etc-podinfo\") pod \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.444560 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-scripts\") pod \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.444617 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/c097cadc-aa5c-43c1-a2b1-1af6739d4086-var-lib-ironic\") pod \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.444661 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-config\") pod \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.444689 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-combined-ca-bundle\") pod \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\" (UID: \"c097cadc-aa5c-43c1-a2b1-1af6739d4086\") " Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.445291 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c097cadc-aa5c-43c1-a2b1-1af6739d4086-var-lib-ironic" (OuterVolumeSpecName: "var-lib-ironic") pod "c097cadc-aa5c-43c1-a2b1-1af6739d4086" (UID: "c097cadc-aa5c-43c1-a2b1-1af6739d4086"). InnerVolumeSpecName "var-lib-ironic". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.445526 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c097cadc-aa5c-43c1-a2b1-1af6739d4086-var-lib-ironic-inspector-dhcp-hostsdir" (OuterVolumeSpecName: "var-lib-ironic-inspector-dhcp-hostsdir") pod "c097cadc-aa5c-43c1-a2b1-1af6739d4086" (UID: "c097cadc-aa5c-43c1-a2b1-1af6739d4086"). InnerVolumeSpecName "var-lib-ironic-inspector-dhcp-hostsdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.451380 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c097cadc-aa5c-43c1-a2b1-1af6739d4086-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "c097cadc-aa5c-43c1-a2b1-1af6739d4086" (UID: "c097cadc-aa5c-43c1-a2b1-1af6739d4086"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.452800 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-config" (OuterVolumeSpecName: "config") pod "c097cadc-aa5c-43c1-a2b1-1af6739d4086" (UID: "c097cadc-aa5c-43c1-a2b1-1af6739d4086"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.453075 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-scripts" (OuterVolumeSpecName: "scripts") pod "c097cadc-aa5c-43c1-a2b1-1af6739d4086" (UID: "c097cadc-aa5c-43c1-a2b1-1af6739d4086"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.454012 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c097cadc-aa5c-43c1-a2b1-1af6739d4086-kube-api-access-vfmhv" (OuterVolumeSpecName: "kube-api-access-vfmhv") pod "c097cadc-aa5c-43c1-a2b1-1af6739d4086" (UID: "c097cadc-aa5c-43c1-a2b1-1af6739d4086"). InnerVolumeSpecName "kube-api-access-vfmhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.483379 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c097cadc-aa5c-43c1-a2b1-1af6739d4086" (UID: "c097cadc-aa5c-43c1-a2b1-1af6739d4086"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.547447 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfmhv\" (UniqueName: \"kubernetes.io/projected/c097cadc-aa5c-43c1-a2b1-1af6739d4086-kube-api-access-vfmhv\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.547492 4945 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/c097cadc-aa5c-43c1-a2b1-1af6739d4086-var-lib-ironic-inspector-dhcp-hostsdir\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.547508 4945 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/c097cadc-aa5c-43c1-a2b1-1af6739d4086-etc-podinfo\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.547522 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.547534 4945 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/c097cadc-aa5c-43c1-a2b1-1af6739d4086-var-lib-ironic\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.547545 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.547556 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c097cadc-aa5c-43c1-a2b1-1af6739d4086-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.762962 4945 scope.go:117] "RemoveContainer" containerID="3603f131d46f535fe562976c5a3e4677fe3730d12b9ddc2cc971f48949614604" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.778018 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd04b8b0-f8ab-4851-8303-c9ab3286bf58" path="/var/lib/kubelet/pods/fd04b8b0-f8ab-4851-8303-c9ab3286bf58/volumes" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.861655 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:34:32 crc kubenswrapper[4945]: W1014 15:34:32.873107 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fdab3dc_df19_4f56_a226_8afb55579e2e.slice/crio-51f7433478f6b9d700a1ecb35580876b1c33e0d42b1c7e7e89e52d465dc62bba WatchSource:0}: Error finding container 51f7433478f6b9d700a1ecb35580876b1c33e0d42b1c7e7e89e52d465dc62bba: Status 404 returned error can't find the container with id 51f7433478f6b9d700a1ecb35580876b1c33e0d42b1c7e7e89e52d465dc62bba Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.898545 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fdab3dc-df19-4f56-a226-8afb55579e2e","Type":"ContainerStarted","Data":"51f7433478f6b9d700a1ecb35580876b1c33e0d42b1c7e7e89e52d465dc62bba"} Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.901393 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.902083 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"c097cadc-aa5c-43c1-a2b1-1af6739d4086","Type":"ContainerDied","Data":"307dba8a52dc1fdeaff859cd54922ea3997bd646f0e256c3ebb53cc7703e2fdf"} Oct 14 15:34:32 crc kubenswrapper[4945]: I1014 15:34:32.902122 4945 scope.go:117] "RemoveContainer" containerID="bd72f628ad6f4818445a894bd5041c2c37945429b577069463c2ab0af9674d95" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.043505 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-0"] Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.052530 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-0"] Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.089793 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-0"] Oct 14 15:34:33 crc kubenswrapper[4945]: E1014 15:34:33.090250 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c097cadc-aa5c-43c1-a2b1-1af6739d4086" containerName="ironic-python-agent-init" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.090263 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c097cadc-aa5c-43c1-a2b1-1af6739d4086" containerName="ironic-python-agent-init" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.090444 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c097cadc-aa5c-43c1-a2b1-1af6739d4086" containerName="ironic-python-agent-init" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.092535 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.096464 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.097365 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-inspector-public-svc" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.097531 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-inspector-internal-svc" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.097657 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.099777 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.185621 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.185663 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.185801 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.185862 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-config\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.185939 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.186106 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5mgl\" (UniqueName: \"kubernetes.io/projected/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-kube-api-access-q5mgl\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.186177 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.186292 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-scripts\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.186327 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.287883 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.288216 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5mgl\" (UniqueName: \"kubernetes.io/projected/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-kube-api-access-q5mgl\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.288251 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.288297 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-scripts\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.288326 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.288381 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.288403 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.288447 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.288482 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-config\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.288895 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.289192 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.293896 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.293890 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.293998 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.294665 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.295087 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-config\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.302800 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-scripts\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.304550 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5mgl\" (UniqueName: \"kubernetes.io/projected/3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1-kube-api-access-q5mgl\") pod \"ironic-inspector-0\" (UID: \"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1\") " pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.429939 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.920846 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.925204 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" event={"ID":"11506640-cea7-421c-878c-c4b449df880f","Type":"ContainerStarted","Data":"a1bc5a258d03c9d726827db38054807029923bde52edef3ed587956e51268af6"} Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.926467 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:34:33 crc kubenswrapper[4945]: I1014 15:34:33.928124 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fdab3dc-df19-4f56-a226-8afb55579e2e","Type":"ContainerStarted","Data":"5f08462e4582c41377a16492373bfb746d8e2f683cec22a0936bc2127a6cd17f"} Oct 14 15:34:33 crc kubenswrapper[4945]: W1014 15:34:33.928210 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e3ee8d8_f91a_40ae_889c_b7e86a13f8f1.slice/crio-b40fdc8e1366e49ad4032fd643c4a5941d12d76281a2f1f348da52c409c2b6b2 WatchSource:0}: Error finding container b40fdc8e1366e49ad4032fd643c4a5941d12d76281a2f1f348da52c409c2b6b2: Status 404 returned error can't find the container with id b40fdc8e1366e49ad4032fd643c4a5941d12d76281a2f1f348da52c409c2b6b2 Oct 14 15:34:34 crc kubenswrapper[4945]: I1014 15:34:34.773584 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c097cadc-aa5c-43c1-a2b1-1af6739d4086" path="/var/lib/kubelet/pods/c097cadc-aa5c-43c1-a2b1-1af6739d4086/volumes" Oct 14 15:34:34 crc kubenswrapper[4945]: I1014 15:34:34.943940 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1","Type":"ContainerDied","Data":"1a6a7f5de25a11e73d0a9a11f3bde5a68ff364373f974fa819e21e926752f79e"} Oct 14 15:34:34 crc kubenswrapper[4945]: I1014 15:34:34.945943 4945 generic.go:334] "Generic (PLEG): container finished" podID="3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1" containerID="1a6a7f5de25a11e73d0a9a11f3bde5a68ff364373f974fa819e21e926752f79e" exitCode=0 Oct 14 15:34:34 crc kubenswrapper[4945]: I1014 15:34:34.946066 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1","Type":"ContainerStarted","Data":"b40fdc8e1366e49ad4032fd643c4a5941d12d76281a2f1f348da52c409c2b6b2"} Oct 14 15:34:34 crc kubenswrapper[4945]: I1014 15:34:34.952186 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fdab3dc-df19-4f56-a226-8afb55579e2e","Type":"ContainerStarted","Data":"318253a2d76677862b0788c33814f2b21b39907b7dc11ef78f43672fb9da3bb8"} Oct 14 15:34:35 crc kubenswrapper[4945]: I1014 15:34:35.965650 4945 generic.go:334] "Generic (PLEG): container finished" podID="042f4fc3-10ec-450e-b3cf-77a11d71a8b7" containerID="62ec4fa672b977907ddecbe53d8a70d1aa572a93de1d45c57b587aea8d8f33c6" exitCode=0 Oct 14 15:34:35 crc kubenswrapper[4945]: I1014 15:34:35.965739 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"042f4fc3-10ec-450e-b3cf-77a11d71a8b7","Type":"ContainerDied","Data":"62ec4fa672b977907ddecbe53d8a70d1aa572a93de1d45c57b587aea8d8f33c6"} Oct 14 15:34:35 crc kubenswrapper[4945]: I1014 15:34:35.971566 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fdab3dc-df19-4f56-a226-8afb55579e2e","Type":"ContainerStarted","Data":"d53cbc3845fcf0ed2b50c66d9e6c5b082638890503f7a44548a519f13d3d076b"} Oct 14 15:34:36 crc kubenswrapper[4945]: I1014 15:34:36.306896 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-neutron-agent-648bd876cc-srkfr" Oct 14 15:34:39 crc kubenswrapper[4945]: I1014 15:34:39.001335 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fdab3dc-df19-4f56-a226-8afb55579e2e","Type":"ContainerStarted","Data":"0844a95c766fb5cbba39086814548a0ac5a6aa9fc4f0da2ba2e65a1b860b9378"} Oct 14 15:34:39 crc kubenswrapper[4945]: I1014 15:34:39.001837 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 15:34:39 crc kubenswrapper[4945]: I1014 15:34:39.022400 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.190973301 podStartE2EDuration="8.022380857s" podCreationTimestamp="2025-10-14 15:34:31 +0000 UTC" firstStartedPulling="2025-10-14 15:34:32.875109978 +0000 UTC m=+1202.859158346" lastFinishedPulling="2025-10-14 15:34:37.706517534 +0000 UTC m=+1207.690565902" observedRunningTime="2025-10-14 15:34:39.01969593 +0000 UTC m=+1209.003744298" watchObservedRunningTime="2025-10-14 15:34:39.022380857 +0000 UTC m=+1209.006429215" Oct 14 15:34:40 crc kubenswrapper[4945]: I1014 15:34:40.273676 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 14 15:34:40 crc kubenswrapper[4945]: I1014 15:34:40.870916 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-b7scm"] Oct 14 15:34:40 crc kubenswrapper[4945]: I1014 15:34:40.873060 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:40 crc kubenswrapper[4945]: I1014 15:34:40.876502 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 14 15:34:40 crc kubenswrapper[4945]: I1014 15:34:40.876571 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 14 15:34:40 crc kubenswrapper[4945]: I1014 15:34:40.884219 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-b7scm"] Oct 14 15:34:40 crc kubenswrapper[4945]: I1014 15:34:40.981119 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-scripts\") pod \"nova-cell0-cell-mapping-b7scm\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:40 crc kubenswrapper[4945]: I1014 15:34:40.981171 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-config-data\") pod \"nova-cell0-cell-mapping-b7scm\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:40 crc kubenswrapper[4945]: I1014 15:34:40.981214 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n79dz\" (UniqueName: \"kubernetes.io/projected/63c34001-2d18-4f96-a9e5-af05eab7f11b-kube-api-access-n79dz\") pod \"nova-cell0-cell-mapping-b7scm\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:40 crc kubenswrapper[4945]: I1014 15:34:40.981319 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-b7scm\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.075732 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.078095 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.085129 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.091193 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-scripts\") pod \"nova-cell0-cell-mapping-b7scm\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.091257 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-config-data\") pod \"nova-cell0-cell-mapping-b7scm\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.091296 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n79dz\" (UniqueName: \"kubernetes.io/projected/63c34001-2d18-4f96-a9e5-af05eab7f11b-kube-api-access-n79dz\") pod \"nova-cell0-cell-mapping-b7scm\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.091437 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-b7scm\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.105563 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-b7scm\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.106201 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-config-data\") pod \"nova-cell0-cell-mapping-b7scm\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.118427 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-scripts\") pod \"nova-cell0-cell-mapping-b7scm\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.119922 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.140099 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n79dz\" (UniqueName: \"kubernetes.io/projected/63c34001-2d18-4f96-a9e5-af05eab7f11b-kube-api-access-n79dz\") pod \"nova-cell0-cell-mapping-b7scm\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.147036 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.148557 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.161219 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.174954 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.197018 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721e6f5e-9053-494a-80f7-127ad3ec46f8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.197114 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/721e6f5e-9053-494a-80f7-127ad3ec46f8-config-data\") pod \"nova-api-0\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.197245 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/721e6f5e-9053-494a-80f7-127ad3ec46f8-logs\") pod \"nova-api-0\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.197275 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs8gj\" (UniqueName: \"kubernetes.io/projected/721e6f5e-9053-494a-80f7-127ad3ec46f8-kube-api-access-vs8gj\") pod \"nova-api-0\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.208972 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.237120 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.238833 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.244349 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.303705 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/721e6f5e-9053-494a-80f7-127ad3ec46f8-config-data\") pod \"nova-api-0\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.303772 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e8370d-1890-4a06-9566-5696dd1f926c-config-data\") pod \"nova-scheduler-0\" (UID: \"48e8370d-1890-4a06-9566-5696dd1f926c\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.303910 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/721e6f5e-9053-494a-80f7-127ad3ec46f8-logs\") pod \"nova-api-0\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.303936 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs8gj\" (UniqueName: \"kubernetes.io/projected/721e6f5e-9053-494a-80f7-127ad3ec46f8-kube-api-access-vs8gj\") pod \"nova-api-0\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.304035 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721e6f5e-9053-494a-80f7-127ad3ec46f8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.304070 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e8370d-1890-4a06-9566-5696dd1f926c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"48e8370d-1890-4a06-9566-5696dd1f926c\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.304092 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt7br\" (UniqueName: \"kubernetes.io/projected/48e8370d-1890-4a06-9566-5696dd1f926c-kube-api-access-wt7br\") pod \"nova-scheduler-0\" (UID: \"48e8370d-1890-4a06-9566-5696dd1f926c\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.304703 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/721e6f5e-9053-494a-80f7-127ad3ec46f8-logs\") pod \"nova-api-0\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.314483 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.321360 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721e6f5e-9053-494a-80f7-127ad3ec46f8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.338356 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/721e6f5e-9053-494a-80f7-127ad3ec46f8-config-data\") pod \"nova-api-0\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.391902 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-hw5kn"] Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.396348 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.399396 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs8gj\" (UniqueName: \"kubernetes.io/projected/721e6f5e-9053-494a-80f7-127ad3ec46f8-kube-api-access-vs8gj\") pod \"nova-api-0\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.407994 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.409843 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-logs\") pod \"nova-metadata-0\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.427954 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt7br\" (UniqueName: \"kubernetes.io/projected/48e8370d-1890-4a06-9566-5696dd1f926c-kube-api-access-wt7br\") pod \"nova-scheduler-0\" (UID: \"48e8370d-1890-4a06-9566-5696dd1f926c\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.428025 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e8370d-1890-4a06-9566-5696dd1f926c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"48e8370d-1890-4a06-9566-5696dd1f926c\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.428190 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e8370d-1890-4a06-9566-5696dd1f926c-config-data\") pod \"nova-scheduler-0\" (UID: \"48e8370d-1890-4a06-9566-5696dd1f926c\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.428235 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-config-data\") pod \"nova-metadata-0\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.428307 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lbx9\" (UniqueName: \"kubernetes.io/projected/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-kube-api-access-6lbx9\") pod \"nova-metadata-0\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.428356 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.438866 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e8370d-1890-4a06-9566-5696dd1f926c-config-data\") pod \"nova-scheduler-0\" (UID: \"48e8370d-1890-4a06-9566-5696dd1f926c\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.453810 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e8370d-1890-4a06-9566-5696dd1f926c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"48e8370d-1890-4a06-9566-5696dd1f926c\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.454784 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt7br\" (UniqueName: \"kubernetes.io/projected/48e8370d-1890-4a06-9566-5696dd1f926c-kube-api-access-wt7br\") pod \"nova-scheduler-0\" (UID: \"48e8370d-1890-4a06-9566-5696dd1f926c\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.455356 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-hw5kn"] Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.488461 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.491457 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.493037 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.523195 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.537236 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gc9b\" (UniqueName: \"kubernetes.io/projected/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-kube-api-access-7gc9b\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.537338 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.537372 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-logs\") pod \"nova-metadata-0\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.537396 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-dns-svc\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.537450 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.537478 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.537554 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-config-data\") pod \"nova-metadata-0\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.537603 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lbx9\" (UniqueName: \"kubernetes.io/projected/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-kube-api-access-6lbx9\") pod \"nova-metadata-0\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.537633 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.537732 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-config\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.538154 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-logs\") pod \"nova-metadata-0\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.544606 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-config-data\") pod \"nova-metadata-0\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.558247 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.561340 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.573814 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lbx9\" (UniqueName: \"kubernetes.io/projected/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-kube-api-access-6lbx9\") pod \"nova-metadata-0\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.639952 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.640041 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-dns-svc\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.640081 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.640101 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.640171 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/314ac186-b037-4e81-953c-78288edadcd4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"314ac186-b037-4e81-953c-78288edadcd4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.640223 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-config\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.640260 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gc9b\" (UniqueName: \"kubernetes.io/projected/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-kube-api-access-7gc9b\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.640276 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n8bq\" (UniqueName: \"kubernetes.io/projected/314ac186-b037-4e81-953c-78288edadcd4-kube-api-access-9n8bq\") pod \"nova-cell1-novncproxy-0\" (UID: \"314ac186-b037-4e81-953c-78288edadcd4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.640293 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/314ac186-b037-4e81-953c-78288edadcd4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"314ac186-b037-4e81-953c-78288edadcd4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.640898 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.641110 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-dns-svc\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.641434 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.641638 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-config\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.642033 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.665012 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gc9b\" (UniqueName: \"kubernetes.io/projected/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-kube-api-access-7gc9b\") pod \"dnsmasq-dns-865f5d856f-hw5kn\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.744804 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n8bq\" (UniqueName: \"kubernetes.io/projected/314ac186-b037-4e81-953c-78288edadcd4-kube-api-access-9n8bq\") pod \"nova-cell1-novncproxy-0\" (UID: \"314ac186-b037-4e81-953c-78288edadcd4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.744853 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/314ac186-b037-4e81-953c-78288edadcd4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"314ac186-b037-4e81-953c-78288edadcd4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.745018 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/314ac186-b037-4e81-953c-78288edadcd4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"314ac186-b037-4e81-953c-78288edadcd4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.748712 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/314ac186-b037-4e81-953c-78288edadcd4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"314ac186-b037-4e81-953c-78288edadcd4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.752212 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/314ac186-b037-4e81-953c-78288edadcd4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"314ac186-b037-4e81-953c-78288edadcd4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.770490 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n8bq\" (UniqueName: \"kubernetes.io/projected/314ac186-b037-4e81-953c-78288edadcd4-kube-api-access-9n8bq\") pod \"nova-cell1-novncproxy-0\" (UID: \"314ac186-b037-4e81-953c-78288edadcd4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.808056 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.818395 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.832432 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.927613 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-b7scm"] Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.970221 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fwjnr"] Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.972318 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.974731 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.978282 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 15:34:41 crc kubenswrapper[4945]: I1014 15:34:41.992589 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fwjnr"] Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.090129 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-b7scm" event={"ID":"63c34001-2d18-4f96-a9e5-af05eab7f11b","Type":"ContainerStarted","Data":"535b8e9332c3b41ffa30452051082ba663b55767ea93e30489077b2efb31bea9"} Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.090815 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.092997 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"721e6f5e-9053-494a-80f7-127ad3ec46f8","Type":"ContainerStarted","Data":"8f180a3d12a48b27e33ef4d5951653422e0ddf2b449ad7e3f51e38383b77e088"} Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.163994 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-scripts\") pod \"nova-cell1-conductor-db-sync-fwjnr\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.164079 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fwjnr\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.164161 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx9j6\" (UniqueName: \"kubernetes.io/projected/10a43a64-535b-4cfe-be95-3f6f06d8eb98-kube-api-access-hx9j6\") pod \"nova-cell1-conductor-db-sync-fwjnr\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.164210 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-config-data\") pod \"nova-cell1-conductor-db-sync-fwjnr\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.176508 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.266157 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-scripts\") pod \"nova-cell1-conductor-db-sync-fwjnr\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.266388 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fwjnr\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.266451 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx9j6\" (UniqueName: \"kubernetes.io/projected/10a43a64-535b-4cfe-be95-3f6f06d8eb98-kube-api-access-hx9j6\") pod \"nova-cell1-conductor-db-sync-fwjnr\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.266474 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-config-data\") pod \"nova-cell1-conductor-db-sync-fwjnr\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.274505 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-scripts\") pod \"nova-cell1-conductor-db-sync-fwjnr\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.274571 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-config-data\") pod \"nova-cell1-conductor-db-sync-fwjnr\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.282484 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fwjnr\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.283477 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx9j6\" (UniqueName: \"kubernetes.io/projected/10a43a64-535b-4cfe-be95-3f6f06d8eb98-kube-api-access-hx9j6\") pod \"nova-cell1-conductor-db-sync-fwjnr\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.313299 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.400711 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-hw5kn"] Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.550378 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 15:34:42 crc kubenswrapper[4945]: W1014 15:34:42.569010 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod314ac186_b037_4e81_953c_78288edadcd4.slice/crio-c6650db9e256dc0912d1a7061aa7d6dd6fdc26d7e8ebba6a07d891f689fc31cf WatchSource:0}: Error finding container c6650db9e256dc0912d1a7061aa7d6dd6fdc26d7e8ebba6a07d891f689fc31cf: Status 404 returned error can't find the container with id c6650db9e256dc0912d1a7061aa7d6dd6fdc26d7e8ebba6a07d891f689fc31cf Oct 14 15:34:42 crc kubenswrapper[4945]: W1014 15:34:42.569844 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0bf28d4_a6aa_40bf_9919_f5353e2a5c71.slice/crio-77bc309d30247504f33d29119edcb31eb29f00bf79f959ef694cbdcaba1f97d3 WatchSource:0}: Error finding container 77bc309d30247504f33d29119edcb31eb29f00bf79f959ef694cbdcaba1f97d3: Status 404 returned error can't find the container with id 77bc309d30247504f33d29119edcb31eb29f00bf79f959ef694cbdcaba1f97d3 Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.614791 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:42 crc kubenswrapper[4945]: I1014 15:34:42.962851 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fwjnr"] Oct 14 15:34:43 crc kubenswrapper[4945]: W1014 15:34:42.999983 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10a43a64_535b_4cfe_be95_3f6f06d8eb98.slice/crio-3ad8e97ed4643e32ca24d1fe9255877e8d9ac43e745812aff2663a4cbd1c796e WatchSource:0}: Error finding container 3ad8e97ed4643e32ca24d1fe9255877e8d9ac43e745812aff2663a4cbd1c796e: Status 404 returned error can't find the container with id 3ad8e97ed4643e32ca24d1fe9255877e8d9ac43e745812aff2663a4cbd1c796e Oct 14 15:34:43 crc kubenswrapper[4945]: I1014 15:34:43.115860 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71","Type":"ContainerStarted","Data":"77bc309d30247504f33d29119edcb31eb29f00bf79f959ef694cbdcaba1f97d3"} Oct 14 15:34:43 crc kubenswrapper[4945]: I1014 15:34:43.117869 4945 generic.go:334] "Generic (PLEG): container finished" podID="fbbb037f-601a-4b65-8d5e-cd0a57e9957b" containerID="dcc670df4f7e3d081db51c05697440d42d5a6f340a7b34d370428e2d0c7dd19c" exitCode=0 Oct 14 15:34:43 crc kubenswrapper[4945]: I1014 15:34:43.118079 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" event={"ID":"fbbb037f-601a-4b65-8d5e-cd0a57e9957b","Type":"ContainerDied","Data":"dcc670df4f7e3d081db51c05697440d42d5a6f340a7b34d370428e2d0c7dd19c"} Oct 14 15:34:43 crc kubenswrapper[4945]: I1014 15:34:43.118097 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" event={"ID":"fbbb037f-601a-4b65-8d5e-cd0a57e9957b","Type":"ContainerStarted","Data":"02f423e833cad5d6ca98a00221bf94092334ecbb5f6bc685f0cb210f7a4e2fec"} Oct 14 15:34:43 crc kubenswrapper[4945]: I1014 15:34:43.126401 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-b7scm" event={"ID":"63c34001-2d18-4f96-a9e5-af05eab7f11b","Type":"ContainerStarted","Data":"d6c931e09a01890871164155e38170b35605e9e2c037ce86c61d29fada94cc7d"} Oct 14 15:34:43 crc kubenswrapper[4945]: I1014 15:34:43.128038 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"48e8370d-1890-4a06-9566-5696dd1f926c","Type":"ContainerStarted","Data":"b4c5eba9ca139888f47d77d0d806d5c7ada06e1f5db7569a7cecf67ad6f0ab49"} Oct 14 15:34:43 crc kubenswrapper[4945]: I1014 15:34:43.129256 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fwjnr" event={"ID":"10a43a64-535b-4cfe-be95-3f6f06d8eb98","Type":"ContainerStarted","Data":"3ad8e97ed4643e32ca24d1fe9255877e8d9ac43e745812aff2663a4cbd1c796e"} Oct 14 15:34:43 crc kubenswrapper[4945]: I1014 15:34:43.130725 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"314ac186-b037-4e81-953c-78288edadcd4","Type":"ContainerStarted","Data":"c6650db9e256dc0912d1a7061aa7d6dd6fdc26d7e8ebba6a07d891f689fc31cf"} Oct 14 15:34:43 crc kubenswrapper[4945]: I1014 15:34:43.170661 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-b7scm" podStartSLOduration=3.170643197 podStartE2EDuration="3.170643197s" podCreationTimestamp="2025-10-14 15:34:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:34:43.15639991 +0000 UTC m=+1213.140448278" watchObservedRunningTime="2025-10-14 15:34:43.170643197 +0000 UTC m=+1213.154691565" Oct 14 15:34:44 crc kubenswrapper[4945]: I1014 15:34:44.150919 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fwjnr" event={"ID":"10a43a64-535b-4cfe-be95-3f6f06d8eb98","Type":"ContainerStarted","Data":"6312c7bc22838dadc44ac2e540aeffc7a384c584fbd94a114b8eec665aa88881"} Oct 14 15:34:44 crc kubenswrapper[4945]: I1014 15:34:44.162066 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" event={"ID":"fbbb037f-601a-4b65-8d5e-cd0a57e9957b","Type":"ContainerStarted","Data":"ce4a3c3cedfd690774e20edda998c949e732f95fa275b6d433dd83fee61c0f76"} Oct 14 15:34:44 crc kubenswrapper[4945]: I1014 15:34:44.162129 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:44 crc kubenswrapper[4945]: I1014 15:34:44.180475 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-fwjnr" podStartSLOduration=3.180440701 podStartE2EDuration="3.180440701s" podCreationTimestamp="2025-10-14 15:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:34:44.166348729 +0000 UTC m=+1214.150397097" watchObservedRunningTime="2025-10-14 15:34:44.180440701 +0000 UTC m=+1214.164489069" Oct 14 15:34:44 crc kubenswrapper[4945]: I1014 15:34:44.205029 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" podStartSLOduration=3.204992512 podStartE2EDuration="3.204992512s" podCreationTimestamp="2025-10-14 15:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:34:44.187349688 +0000 UTC m=+1214.171398056" watchObservedRunningTime="2025-10-14 15:34:44.204992512 +0000 UTC m=+1214.189040900" Oct 14 15:34:45 crc kubenswrapper[4945]: I1014 15:34:45.045228 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:45 crc kubenswrapper[4945]: I1014 15:34:45.060179 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 15:34:46 crc kubenswrapper[4945]: I1014 15:34:46.797412 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:34:46 crc kubenswrapper[4945]: I1014 15:34:46.797689 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.230540 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"042f4fc3-10ec-450e-b3cf-77a11d71a8b7","Type":"ContainerStarted","Data":"463736e1b4ce1f16b5a23d2e524962c60720443527f70ae81106b5f41ac95227"} Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.234560 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1","Type":"ContainerStarted","Data":"94d2422207b7df867e192f02e591e2e08524adc853512a5d8d9e66f0025e19dd"} Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.243031 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71","Type":"ContainerStarted","Data":"eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279"} Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.243078 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71","Type":"ContainerStarted","Data":"05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822"} Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.243214 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" containerName="nova-metadata-log" containerID="cri-o://05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822" gracePeriod=30 Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.243484 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" containerName="nova-metadata-metadata" containerID="cri-o://eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279" gracePeriod=30 Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.252475 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"48e8370d-1890-4a06-9566-5696dd1f926c","Type":"ContainerStarted","Data":"62b703a574a204e6fecc7f134aa1cdc242b83cac536c7568bb49e124b9581c38"} Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.270479 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"721e6f5e-9053-494a-80f7-127ad3ec46f8","Type":"ContainerStarted","Data":"3a272412c4f603eaf36663e4c94d6835d7c4c656a50f940b0ba53bed08eab148"} Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.270528 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"721e6f5e-9053-494a-80f7-127ad3ec46f8","Type":"ContainerStarted","Data":"1eeb489670fa6084db38300f253aac3144a9b047725e66c1648960fbb272d085"} Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.275223 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"314ac186-b037-4e81-953c-78288edadcd4","Type":"ContainerStarted","Data":"aadf692b7b29f8ed5a3f733bb5119e548fb07857bdc3e69d009b43a143e29f62"} Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.275360 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="314ac186-b037-4e81-953c-78288edadcd4" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://aadf692b7b29f8ed5a3f733bb5119e548fb07857bdc3e69d009b43a143e29f62" gracePeriod=30 Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.291833 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.232605346 podStartE2EDuration="8.291806331s" podCreationTimestamp="2025-10-14 15:34:41 +0000 UTC" firstStartedPulling="2025-10-14 15:34:42.193036492 +0000 UTC m=+1212.177084860" lastFinishedPulling="2025-10-14 15:34:48.252237477 +0000 UTC m=+1218.236285845" observedRunningTime="2025-10-14 15:34:49.282862135 +0000 UTC m=+1219.266910503" watchObservedRunningTime="2025-10-14 15:34:49.291806331 +0000 UTC m=+1219.275854689" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.312218 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.658516766 podStartE2EDuration="8.312195833s" podCreationTimestamp="2025-10-14 15:34:41 +0000 UTC" firstStartedPulling="2025-10-14 15:34:42.620343603 +0000 UTC m=+1212.604391971" lastFinishedPulling="2025-10-14 15:34:48.27402267 +0000 UTC m=+1218.258071038" observedRunningTime="2025-10-14 15:34:49.302182437 +0000 UTC m=+1219.286230815" watchObservedRunningTime="2025-10-14 15:34:49.312195833 +0000 UTC m=+1219.296244201" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.378426 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.163702858 podStartE2EDuration="8.378375003s" podCreationTimestamp="2025-10-14 15:34:41 +0000 UTC" firstStartedPulling="2025-10-14 15:34:42.059083027 +0000 UTC m=+1212.043131395" lastFinishedPulling="2025-10-14 15:34:48.273755172 +0000 UTC m=+1218.257803540" observedRunningTime="2025-10-14 15:34:49.360221024 +0000 UTC m=+1219.344269392" watchObservedRunningTime="2025-10-14 15:34:49.378375003 +0000 UTC m=+1219.362423371" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.381011 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.735613588 podStartE2EDuration="8.380998838s" podCreationTimestamp="2025-10-14 15:34:41 +0000 UTC" firstStartedPulling="2025-10-14 15:34:42.620837127 +0000 UTC m=+1212.604885495" lastFinishedPulling="2025-10-14 15:34:48.266222377 +0000 UTC m=+1218.250270745" observedRunningTime="2025-10-14 15:34:49.374478071 +0000 UTC m=+1219.358526439" watchObservedRunningTime="2025-10-14 15:34:49.380998838 +0000 UTC m=+1219.365047206" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.807506 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.884146 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-config-data\") pod \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.884789 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lbx9\" (UniqueName: \"kubernetes.io/projected/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-kube-api-access-6lbx9\") pod \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.884859 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-logs\") pod \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.884925 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-combined-ca-bundle\") pod \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\" (UID: \"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71\") " Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.885209 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-logs" (OuterVolumeSpecName: "logs") pod "e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" (UID: "e0bf28d4-a6aa-40bf-9919-f5353e2a5c71"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.885690 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.888506 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-kube-api-access-6lbx9" (OuterVolumeSpecName: "kube-api-access-6lbx9") pod "e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" (UID: "e0bf28d4-a6aa-40bf-9919-f5353e2a5c71"). InnerVolumeSpecName "kube-api-access-6lbx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.911797 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" (UID: "e0bf28d4-a6aa-40bf-9919-f5353e2a5c71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.933022 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-config-data" (OuterVolumeSpecName: "config-data") pod "e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" (UID: "e0bf28d4-a6aa-40bf-9919-f5353e2a5c71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.986946 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lbx9\" (UniqueName: \"kubernetes.io/projected/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-kube-api-access-6lbx9\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.987162 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:49 crc kubenswrapper[4945]: I1014 15:34:49.987258 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:50 crc kubenswrapper[4945]: E1014 15:34:50.048030 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63c34001_2d18_4f96_a9e5_af05eab7f11b.slice/crio-conmon-d6c931e09a01890871164155e38170b35605e9e2c037ce86c61d29fada94cc7d.scope\": RecentStats: unable to find data in memory cache]" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.312844 4945 generic.go:334] "Generic (PLEG): container finished" podID="63c34001-2d18-4f96-a9e5-af05eab7f11b" containerID="d6c931e09a01890871164155e38170b35605e9e2c037ce86c61d29fada94cc7d" exitCode=0 Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.312928 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-b7scm" event={"ID":"63c34001-2d18-4f96-a9e5-af05eab7f11b","Type":"ContainerDied","Data":"d6c931e09a01890871164155e38170b35605e9e2c037ce86c61d29fada94cc7d"} Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.316705 4945 generic.go:334] "Generic (PLEG): container finished" podID="3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1" containerID="94d2422207b7df867e192f02e591e2e08524adc853512a5d8d9e66f0025e19dd" exitCode=0 Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.316783 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1","Type":"ContainerDied","Data":"94d2422207b7df867e192f02e591e2e08524adc853512a5d8d9e66f0025e19dd"} Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.316814 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1","Type":"ContainerStarted","Data":"1a3b0aec0aefcd814a8ed8258a2eb4c24f83ca85d5e83da452750e41e4b68b7b"} Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.316827 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1","Type":"ContainerStarted","Data":"724c0449f5ea88f6f887d01dbfb5fccf58a84aa865c5c52e77b4d6f7143588bd"} Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.337919 4945 generic.go:334] "Generic (PLEG): container finished" podID="e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" containerID="eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279" exitCode=0 Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.338148 4945 generic.go:334] "Generic (PLEG): container finished" podID="e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" containerID="05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822" exitCode=143 Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.338565 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71","Type":"ContainerDied","Data":"eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279"} Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.338612 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71","Type":"ContainerDied","Data":"05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822"} Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.338623 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0bf28d4-a6aa-40bf-9919-f5353e2a5c71","Type":"ContainerDied","Data":"77bc309d30247504f33d29119edcb31eb29f00bf79f959ef694cbdcaba1f97d3"} Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.338641 4945 scope.go:117] "RemoveContainer" containerID="eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.338813 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.396103 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.408005 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.412854 4945 scope.go:117] "RemoveContainer" containerID="05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.424834 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:50 crc kubenswrapper[4945]: E1014 15:34:50.437483 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" containerName="nova-metadata-metadata" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.437520 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" containerName="nova-metadata-metadata" Oct 14 15:34:50 crc kubenswrapper[4945]: E1014 15:34:50.437532 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" containerName="nova-metadata-log" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.437539 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" containerName="nova-metadata-log" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.437789 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" containerName="nova-metadata-log" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.437810 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" containerName="nova-metadata-metadata" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.438761 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.438849 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.442845 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.443466 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.463995 4945 scope.go:117] "RemoveContainer" containerID="eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279" Oct 14 15:34:50 crc kubenswrapper[4945]: E1014 15:34:50.464393 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279\": container with ID starting with eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279 not found: ID does not exist" containerID="eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.464504 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279"} err="failed to get container status \"eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279\": rpc error: code = NotFound desc = could not find container \"eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279\": container with ID starting with eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279 not found: ID does not exist" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.464597 4945 scope.go:117] "RemoveContainer" containerID="05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822" Oct 14 15:34:50 crc kubenswrapper[4945]: E1014 15:34:50.465444 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822\": container with ID starting with 05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822 not found: ID does not exist" containerID="05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.465477 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822"} err="failed to get container status \"05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822\": rpc error: code = NotFound desc = could not find container \"05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822\": container with ID starting with 05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822 not found: ID does not exist" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.465504 4945 scope.go:117] "RemoveContainer" containerID="eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.465750 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279"} err="failed to get container status \"eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279\": rpc error: code = NotFound desc = could not find container \"eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279\": container with ID starting with eaf1c696d887e5a2e20244af0837506b8d5f2b662662f1dba770e5679b0e1279 not found: ID does not exist" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.465831 4945 scope.go:117] "RemoveContainer" containerID="05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.466143 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822"} err="failed to get container status \"05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822\": rpc error: code = NotFound desc = could not find container \"05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822\": container with ID starting with 05f30624e977383134388afe55ef10965c3bf1bd0a0fd779347bfe7a6b0f4822 not found: ID does not exist" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.600049 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.600117 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.600315 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnwtd\" (UniqueName: \"kubernetes.io/projected/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-kube-api-access-lnwtd\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.600357 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-logs\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.600400 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-config-data\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.701908 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnwtd\" (UniqueName: \"kubernetes.io/projected/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-kube-api-access-lnwtd\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.701971 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-logs\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.702031 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-config-data\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.702086 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.702136 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.703238 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-logs\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.708772 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.714572 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-config-data\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.731005 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnwtd\" (UniqueName: \"kubernetes.io/projected/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-kube-api-access-lnwtd\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.743533 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.762446 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:34:50 crc kubenswrapper[4945]: I1014 15:34:50.775008 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0bf28d4-a6aa-40bf-9919-f5353e2a5c71" path="/var/lib/kubelet/pods/e0bf28d4-a6aa-40bf-9919-f5353e2a5c71/volumes" Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.193996 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:51 crc kubenswrapper[4945]: W1014 15:34:51.202718 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c809d1b_7252_4fb1_b2cf_1e549158b0d3.slice/crio-8aeb6c5b8597f354c708d517ac2188f3a243b5101e5d3bc3f94986c26e35ecdc WatchSource:0}: Error finding container 8aeb6c5b8597f354c708d517ac2188f3a243b5101e5d3bc3f94986c26e35ecdc: Status 404 returned error can't find the container with id 8aeb6c5b8597f354c708d517ac2188f3a243b5101e5d3bc3f94986c26e35ecdc Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.358646 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3c809d1b-7252-4fb1-b2cf-1e549158b0d3","Type":"ContainerStarted","Data":"8aeb6c5b8597f354c708d517ac2188f3a243b5101e5d3bc3f94986c26e35ecdc"} Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.368808 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1","Type":"ContainerStarted","Data":"ede167c11a305ef3c22d78c63598e77b52b840f87a8612db94af6e4f8d02d9e8"} Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.410752 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.410806 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.563050 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.563102 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.624740 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.801060 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.820678 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.833266 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.933043 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-49n26"] Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.933278 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" podUID="ea523baf-11e5-412e-b5ea-84b3dff1bcb3" containerName="dnsmasq-dns" containerID="cri-o://0e69d3d567d012f6328607ddd7ea353000e2d899addbff317eed50910ffc2ad6" gracePeriod=10 Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.938137 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-config-data\") pod \"63c34001-2d18-4f96-a9e5-af05eab7f11b\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.938180 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-scripts\") pod \"63c34001-2d18-4f96-a9e5-af05eab7f11b\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.938289 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n79dz\" (UniqueName: \"kubernetes.io/projected/63c34001-2d18-4f96-a9e5-af05eab7f11b-kube-api-access-n79dz\") pod \"63c34001-2d18-4f96-a9e5-af05eab7f11b\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.938327 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-combined-ca-bundle\") pod \"63c34001-2d18-4f96-a9e5-af05eab7f11b\" (UID: \"63c34001-2d18-4f96-a9e5-af05eab7f11b\") " Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.951629 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63c34001-2d18-4f96-a9e5-af05eab7f11b-kube-api-access-n79dz" (OuterVolumeSpecName: "kube-api-access-n79dz") pod "63c34001-2d18-4f96-a9e5-af05eab7f11b" (UID: "63c34001-2d18-4f96-a9e5-af05eab7f11b"). InnerVolumeSpecName "kube-api-access-n79dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.952155 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-scripts" (OuterVolumeSpecName: "scripts") pod "63c34001-2d18-4f96-a9e5-af05eab7f11b" (UID: "63c34001-2d18-4f96-a9e5-af05eab7f11b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.986508 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63c34001-2d18-4f96-a9e5-af05eab7f11b" (UID: "63c34001-2d18-4f96-a9e5-af05eab7f11b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:51 crc kubenswrapper[4945]: I1014 15:34:51.988410 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-config-data" (OuterVolumeSpecName: "config-data") pod "63c34001-2d18-4f96-a9e5-af05eab7f11b" (UID: "63c34001-2d18-4f96-a9e5-af05eab7f11b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.040885 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.040923 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.040936 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n79dz\" (UniqueName: \"kubernetes.io/projected/63c34001-2d18-4f96-a9e5-af05eab7f11b-kube-api-access-n79dz\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.040950 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c34001-2d18-4f96-a9e5-af05eab7f11b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.382622 4945 generic.go:334] "Generic (PLEG): container finished" podID="ea523baf-11e5-412e-b5ea-84b3dff1bcb3" containerID="0e69d3d567d012f6328607ddd7ea353000e2d899addbff317eed50910ffc2ad6" exitCode=0 Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.382980 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" event={"ID":"ea523baf-11e5-412e-b5ea-84b3dff1bcb3","Type":"ContainerDied","Data":"0e69d3d567d012f6328607ddd7ea353000e2d899addbff317eed50910ffc2ad6"} Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.383013 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" event={"ID":"ea523baf-11e5-412e-b5ea-84b3dff1bcb3","Type":"ContainerDied","Data":"0807b194ee7d42fc4abd8bc8cf071e1a6fe3697edc5ca0ca6590095fd58b72e9"} Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.383027 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0807b194ee7d42fc4abd8bc8cf071e1a6fe3697edc5ca0ca6590095fd58b72e9" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.385164 4945 generic.go:334] "Generic (PLEG): container finished" podID="10a43a64-535b-4cfe-be95-3f6f06d8eb98" containerID="6312c7bc22838dadc44ac2e540aeffc7a384c584fbd94a114b8eec665aa88881" exitCode=0 Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.385230 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fwjnr" event={"ID":"10a43a64-535b-4cfe-be95-3f6f06d8eb98","Type":"ContainerDied","Data":"6312c7bc22838dadc44ac2e540aeffc7a384c584fbd94a114b8eec665aa88881"} Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.388483 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-b7scm" event={"ID":"63c34001-2d18-4f96-a9e5-af05eab7f11b","Type":"ContainerDied","Data":"535b8e9332c3b41ffa30452051082ba663b55767ea93e30489077b2efb31bea9"} Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.388519 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="535b8e9332c3b41ffa30452051082ba663b55767ea93e30489077b2efb31bea9" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.388572 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-b7scm" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.402845 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1","Type":"ContainerStarted","Data":"4e567a1489a04aeed9e93edae2afc96e6768f32e057e12c8f44da9a614d3cad0"} Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.404121 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.410840 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3c809d1b-7252-4fb1-b2cf-1e549158b0d3","Type":"ContainerStarted","Data":"a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560"} Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.410899 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3c809d1b-7252-4fb1-b2cf-1e549158b0d3","Type":"ContainerStarted","Data":"9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33"} Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.450214 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-inspector-0" podStartSLOduration=6.122445243 podStartE2EDuration="19.450190635s" podCreationTimestamp="2025-10-14 15:34:33 +0000 UTC" firstStartedPulling="2025-10-14 15:34:34.947042249 +0000 UTC m=+1204.931090617" lastFinishedPulling="2025-10-14 15:34:48.274787651 +0000 UTC m=+1218.258836009" observedRunningTime="2025-10-14 15:34:52.43774312 +0000 UTC m=+1222.421791498" watchObservedRunningTime="2025-10-14 15:34:52.450190635 +0000 UTC m=+1222.434239003" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.452193 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.455559 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.465908 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.4658606929999998 podStartE2EDuration="2.465860693s" podCreationTimestamp="2025-10-14 15:34:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:34:52.459362907 +0000 UTC m=+1222.443411285" watchObservedRunningTime="2025-10-14 15:34:52.465860693 +0000 UTC m=+1222.449909071" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.495201 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="721e6f5e-9053-494a-80f7-127ad3ec46f8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.495213 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="721e6f5e-9053-494a-80f7-127ad3ec46f8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.549378 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdjc8\" (UniqueName: \"kubernetes.io/projected/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-kube-api-access-mdjc8\") pod \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.549443 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-ovsdbserver-sb\") pod \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.549515 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-dns-swift-storage-0\") pod \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.549579 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-dns-svc\") pod \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.549640 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-ovsdbserver-nb\") pod \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.549669 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-config\") pod \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\" (UID: \"ea523baf-11e5-412e-b5ea-84b3dff1bcb3\") " Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.561462 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-kube-api-access-mdjc8" (OuterVolumeSpecName: "kube-api-access-mdjc8") pod "ea523baf-11e5-412e-b5ea-84b3dff1bcb3" (UID: "ea523baf-11e5-412e-b5ea-84b3dff1bcb3"). InnerVolumeSpecName "kube-api-access-mdjc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.600805 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.601099 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="721e6f5e-9053-494a-80f7-127ad3ec46f8" containerName="nova-api-log" containerID="cri-o://1eeb489670fa6084db38300f253aac3144a9b047725e66c1648960fbb272d085" gracePeriod=30 Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.601199 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="721e6f5e-9053-494a-80f7-127ad3ec46f8" containerName="nova-api-api" containerID="cri-o://3a272412c4f603eaf36663e4c94d6835d7c4c656a50f940b0ba53bed08eab148" gracePeriod=30 Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.614218 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.654990 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdjc8\" (UniqueName: \"kubernetes.io/projected/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-kube-api-access-mdjc8\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.669146 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea523baf-11e5-412e-b5ea-84b3dff1bcb3" (UID: "ea523baf-11e5-412e-b5ea-84b3dff1bcb3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.672002 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ea523baf-11e5-412e-b5ea-84b3dff1bcb3" (UID: "ea523baf-11e5-412e-b5ea-84b3dff1bcb3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.685020 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea523baf-11e5-412e-b5ea-84b3dff1bcb3" (UID: "ea523baf-11e5-412e-b5ea-84b3dff1bcb3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.705583 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-config" (OuterVolumeSpecName: "config") pod "ea523baf-11e5-412e-b5ea-84b3dff1bcb3" (UID: "ea523baf-11e5-412e-b5ea-84b3dff1bcb3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.711216 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea523baf-11e5-412e-b5ea-84b3dff1bcb3" (UID: "ea523baf-11e5-412e-b5ea-84b3dff1bcb3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.756192 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.756225 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.756239 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.756252 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.756261 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea523baf-11e5-412e-b5ea-84b3dff1bcb3-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:52 crc kubenswrapper[4945]: I1014 15:34:52.944370 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.425680 4945 generic.go:334] "Generic (PLEG): container finished" podID="721e6f5e-9053-494a-80f7-127ad3ec46f8" containerID="1eeb489670fa6084db38300f253aac3144a9b047725e66c1648960fbb272d085" exitCode=143 Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.425728 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"721e6f5e-9053-494a-80f7-127ad3ec46f8","Type":"ContainerDied","Data":"1eeb489670fa6084db38300f253aac3144a9b047725e66c1648960fbb272d085"} Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.425792 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-49n26" Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.430245 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.434398 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-inspector-0" Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.434420 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-inspector-0" Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.434431 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.448752 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/ironic-inspector-0" podUID="3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1" containerName="ironic-inspector" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.448772 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/ironic-inspector-0" podUID="3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1" containerName="ironic-inspector-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.455130 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-49n26"] Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.462814 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-49n26"] Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.851297 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.978004 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-config-data\") pod \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.978076 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hx9j6\" (UniqueName: \"kubernetes.io/projected/10a43a64-535b-4cfe-be95-3f6f06d8eb98-kube-api-access-hx9j6\") pod \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.978100 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-scripts\") pod \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.978149 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-combined-ca-bundle\") pod \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\" (UID: \"10a43a64-535b-4cfe-be95-3f6f06d8eb98\") " Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.982943 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10a43a64-535b-4cfe-be95-3f6f06d8eb98-kube-api-access-hx9j6" (OuterVolumeSpecName: "kube-api-access-hx9j6") pod "10a43a64-535b-4cfe-be95-3f6f06d8eb98" (UID: "10a43a64-535b-4cfe-be95-3f6f06d8eb98"). InnerVolumeSpecName "kube-api-access-hx9j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:53 crc kubenswrapper[4945]: I1014 15:34:53.988484 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-scripts" (OuterVolumeSpecName: "scripts") pod "10a43a64-535b-4cfe-be95-3f6f06d8eb98" (UID: "10a43a64-535b-4cfe-be95-3f6f06d8eb98"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.009758 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10a43a64-535b-4cfe-be95-3f6f06d8eb98" (UID: "10a43a64-535b-4cfe-be95-3f6f06d8eb98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.023014 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-config-data" (OuterVolumeSpecName: "config-data") pod "10a43a64-535b-4cfe-be95-3f6f06d8eb98" (UID: "10a43a64-535b-4cfe-be95-3f6f06d8eb98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.080961 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.081014 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hx9j6\" (UniqueName: \"kubernetes.io/projected/10a43a64-535b-4cfe-be95-3f6f06d8eb98-kube-api-access-hx9j6\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.081113 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.081129 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a43a64-535b-4cfe-be95-3f6f06d8eb98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.442714 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3c809d1b-7252-4fb1-b2cf-1e549158b0d3" containerName="nova-metadata-log" containerID="cri-o://9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33" gracePeriod=30 Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.443406 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fwjnr" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.444040 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fwjnr" event={"ID":"10a43a64-535b-4cfe-be95-3f6f06d8eb98","Type":"ContainerDied","Data":"3ad8e97ed4643e32ca24d1fe9255877e8d9ac43e745812aff2663a4cbd1c796e"} Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.444077 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ad8e97ed4643e32ca24d1fe9255877e8d9ac43e745812aff2663a4cbd1c796e" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.444213 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3c809d1b-7252-4fb1-b2cf-1e549158b0d3" containerName="nova-metadata-metadata" containerID="cri-o://a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560" gracePeriod=30 Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.444237 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="48e8370d-1890-4a06-9566-5696dd1f926c" containerName="nova-scheduler-scheduler" containerID="cri-o://62b703a574a204e6fecc7f134aa1cdc242b83cac536c7568bb49e124b9581c38" gracePeriod=30 Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.476293 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 15:34:54 crc kubenswrapper[4945]: E1014 15:34:54.476636 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea523baf-11e5-412e-b5ea-84b3dff1bcb3" containerName="init" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.476653 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea523baf-11e5-412e-b5ea-84b3dff1bcb3" containerName="init" Oct 14 15:34:54 crc kubenswrapper[4945]: E1014 15:34:54.476676 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a43a64-535b-4cfe-be95-3f6f06d8eb98" containerName="nova-cell1-conductor-db-sync" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.476683 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a43a64-535b-4cfe-be95-3f6f06d8eb98" containerName="nova-cell1-conductor-db-sync" Oct 14 15:34:54 crc kubenswrapper[4945]: E1014 15:34:54.476694 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea523baf-11e5-412e-b5ea-84b3dff1bcb3" containerName="dnsmasq-dns" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.476700 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea523baf-11e5-412e-b5ea-84b3dff1bcb3" containerName="dnsmasq-dns" Oct 14 15:34:54 crc kubenswrapper[4945]: E1014 15:34:54.476716 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c34001-2d18-4f96-a9e5-af05eab7f11b" containerName="nova-manage" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.476721 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c34001-2d18-4f96-a9e5-af05eab7f11b" containerName="nova-manage" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.476906 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea523baf-11e5-412e-b5ea-84b3dff1bcb3" containerName="dnsmasq-dns" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.476924 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="63c34001-2d18-4f96-a9e5-af05eab7f11b" containerName="nova-manage" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.476937 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a43a64-535b-4cfe-be95-3f6f06d8eb98" containerName="nova-cell1-conductor-db-sync" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.477495 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.481672 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.492700 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.591664 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaedddf9-b122-48e8-9dfc-e3525faaf9d3-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"aaedddf9-b122-48e8-9dfc-e3525faaf9d3\") " pod="openstack/nova-cell1-conductor-0" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.592855 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55tn2\" (UniqueName: \"kubernetes.io/projected/aaedddf9-b122-48e8-9dfc-e3525faaf9d3-kube-api-access-55tn2\") pod \"nova-cell1-conductor-0\" (UID: \"aaedddf9-b122-48e8-9dfc-e3525faaf9d3\") " pod="openstack/nova-cell1-conductor-0" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.592993 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaedddf9-b122-48e8-9dfc-e3525faaf9d3-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"aaedddf9-b122-48e8-9dfc-e3525faaf9d3\") " pod="openstack/nova-cell1-conductor-0" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.694593 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaedddf9-b122-48e8-9dfc-e3525faaf9d3-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"aaedddf9-b122-48e8-9dfc-e3525faaf9d3\") " pod="openstack/nova-cell1-conductor-0" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.694803 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55tn2\" (UniqueName: \"kubernetes.io/projected/aaedddf9-b122-48e8-9dfc-e3525faaf9d3-kube-api-access-55tn2\") pod \"nova-cell1-conductor-0\" (UID: \"aaedddf9-b122-48e8-9dfc-e3525faaf9d3\") " pod="openstack/nova-cell1-conductor-0" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.694857 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaedddf9-b122-48e8-9dfc-e3525faaf9d3-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"aaedddf9-b122-48e8-9dfc-e3525faaf9d3\") " pod="openstack/nova-cell1-conductor-0" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.699697 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaedddf9-b122-48e8-9dfc-e3525faaf9d3-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"aaedddf9-b122-48e8-9dfc-e3525faaf9d3\") " pod="openstack/nova-cell1-conductor-0" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.701328 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaedddf9-b122-48e8-9dfc-e3525faaf9d3-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"aaedddf9-b122-48e8-9dfc-e3525faaf9d3\") " pod="openstack/nova-cell1-conductor-0" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.714503 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55tn2\" (UniqueName: \"kubernetes.io/projected/aaedddf9-b122-48e8-9dfc-e3525faaf9d3-kube-api-access-55tn2\") pod \"nova-cell1-conductor-0\" (UID: \"aaedddf9-b122-48e8-9dfc-e3525faaf9d3\") " pod="openstack/nova-cell1-conductor-0" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.777496 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea523baf-11e5-412e-b5ea-84b3dff1bcb3" path="/var/lib/kubelet/pods/ea523baf-11e5-412e-b5ea-84b3dff1bcb3/volumes" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.795948 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 15:34:54 crc kubenswrapper[4945]: I1014 15:34:54.945296 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.101387 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-combined-ca-bundle\") pod \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.101595 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-config-data\") pod \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.101644 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnwtd\" (UniqueName: \"kubernetes.io/projected/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-kube-api-access-lnwtd\") pod \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.102257 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-nova-metadata-tls-certs\") pod \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.102667 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-logs\") pod \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\" (UID: \"3c809d1b-7252-4fb1-b2cf-1e549158b0d3\") " Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.103067 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-logs" (OuterVolumeSpecName: "logs") pod "3c809d1b-7252-4fb1-b2cf-1e549158b0d3" (UID: "3c809d1b-7252-4fb1-b2cf-1e549158b0d3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.103512 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.106143 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-kube-api-access-lnwtd" (OuterVolumeSpecName: "kube-api-access-lnwtd") pod "3c809d1b-7252-4fb1-b2cf-1e549158b0d3" (UID: "3c809d1b-7252-4fb1-b2cf-1e549158b0d3"). InnerVolumeSpecName "kube-api-access-lnwtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.127836 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-config-data" (OuterVolumeSpecName: "config-data") pod "3c809d1b-7252-4fb1-b2cf-1e549158b0d3" (UID: "3c809d1b-7252-4fb1-b2cf-1e549158b0d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.130345 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c809d1b-7252-4fb1-b2cf-1e549158b0d3" (UID: "3c809d1b-7252-4fb1-b2cf-1e549158b0d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.150390 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3c809d1b-7252-4fb1-b2cf-1e549158b0d3" (UID: "3c809d1b-7252-4fb1-b2cf-1e549158b0d3"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.205027 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.205058 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.205068 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnwtd\" (UniqueName: \"kubernetes.io/projected/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-kube-api-access-lnwtd\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.205079 4945 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c809d1b-7252-4fb1-b2cf-1e549158b0d3-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.264111 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 15:34:55 crc kubenswrapper[4945]: W1014 15:34:55.266901 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaaedddf9_b122_48e8_9dfc_e3525faaf9d3.slice/crio-629717cbc47136a72507022eab4e76400fa28d8e7a6665fad7fa6a99f7291e1f WatchSource:0}: Error finding container 629717cbc47136a72507022eab4e76400fa28d8e7a6665fad7fa6a99f7291e1f: Status 404 returned error can't find the container with id 629717cbc47136a72507022eab4e76400fa28d8e7a6665fad7fa6a99f7291e1f Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.452352 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"aaedddf9-b122-48e8-9dfc-e3525faaf9d3","Type":"ContainerStarted","Data":"7c02e08a4e6d7009bb74e36a51bc3d2e348a27f13f6552e2b97e007d59c17936"} Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.452429 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"aaedddf9-b122-48e8-9dfc-e3525faaf9d3","Type":"ContainerStarted","Data":"629717cbc47136a72507022eab4e76400fa28d8e7a6665fad7fa6a99f7291e1f"} Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.452480 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.456024 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c809d1b-7252-4fb1-b2cf-1e549158b0d3" containerID="a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560" exitCode=0 Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.456052 4945 generic.go:334] "Generic (PLEG): container finished" podID="3c809d1b-7252-4fb1-b2cf-1e549158b0d3" containerID="9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33" exitCode=143 Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.456092 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.456123 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3c809d1b-7252-4fb1-b2cf-1e549158b0d3","Type":"ContainerDied","Data":"a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560"} Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.456168 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3c809d1b-7252-4fb1-b2cf-1e549158b0d3","Type":"ContainerDied","Data":"9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33"} Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.456181 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3c809d1b-7252-4fb1-b2cf-1e549158b0d3","Type":"ContainerDied","Data":"8aeb6c5b8597f354c708d517ac2188f3a243b5101e5d3bc3f94986c26e35ecdc"} Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.456201 4945 scope.go:117] "RemoveContainer" containerID="a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.457518 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.471703 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.471576399 podStartE2EDuration="1.471576399s" podCreationTimestamp="2025-10-14 15:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:34:55.468308905 +0000 UTC m=+1225.452357273" watchObservedRunningTime="2025-10-14 15:34:55.471576399 +0000 UTC m=+1225.455624777" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.508607 4945 scope.go:117] "RemoveContainer" containerID="9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.540684 4945 scope.go:117] "RemoveContainer" containerID="a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560" Oct 14 15:34:55 crc kubenswrapper[4945]: E1014 15:34:55.543419 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560\": container with ID starting with a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560 not found: ID does not exist" containerID="a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.543479 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560"} err="failed to get container status \"a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560\": rpc error: code = NotFound desc = could not find container \"a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560\": container with ID starting with a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560 not found: ID does not exist" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.543514 4945 scope.go:117] "RemoveContainer" containerID="9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33" Oct 14 15:34:55 crc kubenswrapper[4945]: E1014 15:34:55.543935 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33\": container with ID starting with 9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33 not found: ID does not exist" containerID="9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.543963 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33"} err="failed to get container status \"9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33\": rpc error: code = NotFound desc = could not find container \"9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33\": container with ID starting with 9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33 not found: ID does not exist" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.543986 4945 scope.go:117] "RemoveContainer" containerID="a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.547305 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.547951 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560"} err="failed to get container status \"a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560\": rpc error: code = NotFound desc = could not find container \"a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560\": container with ID starting with a2d3794d613c58f4258b011c424b12a9f9c4d9dd57cdfc8d090de78b3d68d560 not found: ID does not exist" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.547977 4945 scope.go:117] "RemoveContainer" containerID="9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.552881 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33"} err="failed to get container status \"9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33\": rpc error: code = NotFound desc = could not find container \"9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33\": container with ID starting with 9fc53cac6125f2e7f5ad2e8212f058e048ba012a8b5d8eb81570659eb3942e33 not found: ID does not exist" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.559499 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.569779 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:55 crc kubenswrapper[4945]: E1014 15:34:55.570325 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c809d1b-7252-4fb1-b2cf-1e549158b0d3" containerName="nova-metadata-metadata" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.570350 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c809d1b-7252-4fb1-b2cf-1e549158b0d3" containerName="nova-metadata-metadata" Oct 14 15:34:55 crc kubenswrapper[4945]: E1014 15:34:55.570396 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c809d1b-7252-4fb1-b2cf-1e549158b0d3" containerName="nova-metadata-log" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.570406 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c809d1b-7252-4fb1-b2cf-1e549158b0d3" containerName="nova-metadata-log" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.570622 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c809d1b-7252-4fb1-b2cf-1e549158b0d3" containerName="nova-metadata-log" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.570665 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c809d1b-7252-4fb1-b2cf-1e549158b0d3" containerName="nova-metadata-metadata" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.571819 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.574369 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.574669 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.581550 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.713936 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.713973 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-config-data\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.714117 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.714205 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e25b244-f49c-43e0-9cdd-395f0906c15c-logs\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.714237 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24xhb\" (UniqueName: \"kubernetes.io/projected/5e25b244-f49c-43e0-9cdd-395f0906c15c-kube-api-access-24xhb\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.815763 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e25b244-f49c-43e0-9cdd-395f0906c15c-logs\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.816167 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24xhb\" (UniqueName: \"kubernetes.io/projected/5e25b244-f49c-43e0-9cdd-395f0906c15c-kube-api-access-24xhb\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.816240 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.816267 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-config-data\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.816395 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.817465 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e25b244-f49c-43e0-9cdd-395f0906c15c-logs\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.822676 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.822982 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-config-data\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.823466 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.837539 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24xhb\" (UniqueName: \"kubernetes.io/projected/5e25b244-f49c-43e0-9cdd-395f0906c15c-kube-api-access-24xhb\") pod \"nova-metadata-0\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " pod="openstack/nova-metadata-0" Oct 14 15:34:55 crc kubenswrapper[4945]: I1014 15:34:55.926681 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:34:56 crc kubenswrapper[4945]: I1014 15:34:56.386197 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:34:56 crc kubenswrapper[4945]: W1014 15:34:56.467371 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e25b244_f49c_43e0_9cdd_395f0906c15c.slice/crio-2d67eb56b6cfe962edfb78bb95b9ed5aec0e93830364aed13e86d0dbddf6fb19 WatchSource:0}: Error finding container 2d67eb56b6cfe962edfb78bb95b9ed5aec0e93830364aed13e86d0dbddf6fb19: Status 404 returned error can't find the container with id 2d67eb56b6cfe962edfb78bb95b9ed5aec0e93830364aed13e86d0dbddf6fb19 Oct 14 15:34:56 crc kubenswrapper[4945]: I1014 15:34:56.475220 4945 generic.go:334] "Generic (PLEG): container finished" podID="48e8370d-1890-4a06-9566-5696dd1f926c" containerID="62b703a574a204e6fecc7f134aa1cdc242b83cac536c7568bb49e124b9581c38" exitCode=0 Oct 14 15:34:56 crc kubenswrapper[4945]: I1014 15:34:56.476503 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"48e8370d-1890-4a06-9566-5696dd1f926c","Type":"ContainerDied","Data":"62b703a574a204e6fecc7f134aa1cdc242b83cac536c7568bb49e124b9581c38"} Oct 14 15:34:56 crc kubenswrapper[4945]: E1014 15:34:56.564144 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 62b703a574a204e6fecc7f134aa1cdc242b83cac536c7568bb49e124b9581c38 is running failed: container process not found" containerID="62b703a574a204e6fecc7f134aa1cdc242b83cac536c7568bb49e124b9581c38" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 15:34:56 crc kubenswrapper[4945]: E1014 15:34:56.564587 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 62b703a574a204e6fecc7f134aa1cdc242b83cac536c7568bb49e124b9581c38 is running failed: container process not found" containerID="62b703a574a204e6fecc7f134aa1cdc242b83cac536c7568bb49e124b9581c38" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 15:34:56 crc kubenswrapper[4945]: E1014 15:34:56.565531 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 62b703a574a204e6fecc7f134aa1cdc242b83cac536c7568bb49e124b9581c38 is running failed: container process not found" containerID="62b703a574a204e6fecc7f134aa1cdc242b83cac536c7568bb49e124b9581c38" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 15:34:56 crc kubenswrapper[4945]: E1014 15:34:56.565563 4945 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 62b703a574a204e6fecc7f134aa1cdc242b83cac536c7568bb49e124b9581c38 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="48e8370d-1890-4a06-9566-5696dd1f926c" containerName="nova-scheduler-scheduler" Oct 14 15:34:56 crc kubenswrapper[4945]: I1014 15:34:56.801510 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c809d1b-7252-4fb1-b2cf-1e549158b0d3" path="/var/lib/kubelet/pods/3c809d1b-7252-4fb1-b2cf-1e549158b0d3/volumes" Oct 14 15:34:56 crc kubenswrapper[4945]: I1014 15:34:56.880753 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 15:34:56 crc kubenswrapper[4945]: I1014 15:34:56.938712 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e8370d-1890-4a06-9566-5696dd1f926c-combined-ca-bundle\") pod \"48e8370d-1890-4a06-9566-5696dd1f926c\" (UID: \"48e8370d-1890-4a06-9566-5696dd1f926c\") " Oct 14 15:34:56 crc kubenswrapper[4945]: I1014 15:34:56.938865 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt7br\" (UniqueName: \"kubernetes.io/projected/48e8370d-1890-4a06-9566-5696dd1f926c-kube-api-access-wt7br\") pod \"48e8370d-1890-4a06-9566-5696dd1f926c\" (UID: \"48e8370d-1890-4a06-9566-5696dd1f926c\") " Oct 14 15:34:56 crc kubenswrapper[4945]: I1014 15:34:56.938963 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e8370d-1890-4a06-9566-5696dd1f926c-config-data\") pod \"48e8370d-1890-4a06-9566-5696dd1f926c\" (UID: \"48e8370d-1890-4a06-9566-5696dd1f926c\") " Oct 14 15:34:56 crc kubenswrapper[4945]: I1014 15:34:56.957001 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48e8370d-1890-4a06-9566-5696dd1f926c-kube-api-access-wt7br" (OuterVolumeSpecName: "kube-api-access-wt7br") pod "48e8370d-1890-4a06-9566-5696dd1f926c" (UID: "48e8370d-1890-4a06-9566-5696dd1f926c"). InnerVolumeSpecName "kube-api-access-wt7br". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.003128 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e8370d-1890-4a06-9566-5696dd1f926c-config-data" (OuterVolumeSpecName: "config-data") pod "48e8370d-1890-4a06-9566-5696dd1f926c" (UID: "48e8370d-1890-4a06-9566-5696dd1f926c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.017051 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e8370d-1890-4a06-9566-5696dd1f926c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48e8370d-1890-4a06-9566-5696dd1f926c" (UID: "48e8370d-1890-4a06-9566-5696dd1f926c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.041511 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e8370d-1890-4a06-9566-5696dd1f926c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.041561 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt7br\" (UniqueName: \"kubernetes.io/projected/48e8370d-1890-4a06-9566-5696dd1f926c-kube-api-access-wt7br\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.041579 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e8370d-1890-4a06-9566-5696dd1f926c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.485909 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e25b244-f49c-43e0-9cdd-395f0906c15c","Type":"ContainerStarted","Data":"837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622"} Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.486813 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e25b244-f49c-43e0-9cdd-395f0906c15c","Type":"ContainerStarted","Data":"6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64"} Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.486903 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e25b244-f49c-43e0-9cdd-395f0906c15c","Type":"ContainerStarted","Data":"2d67eb56b6cfe962edfb78bb95b9ed5aec0e93830364aed13e86d0dbddf6fb19"} Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.487760 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"48e8370d-1890-4a06-9566-5696dd1f926c","Type":"ContainerDied","Data":"b4c5eba9ca139888f47d77d0d806d5c7ada06e1f5db7569a7cecf67ad6f0ab49"} Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.487811 4945 scope.go:117] "RemoveContainer" containerID="62b703a574a204e6fecc7f134aa1cdc242b83cac536c7568bb49e124b9581c38" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.487772 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.530625 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.530599503 podStartE2EDuration="2.530599503s" podCreationTimestamp="2025-10-14 15:34:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:34:57.526911858 +0000 UTC m=+1227.510960236" watchObservedRunningTime="2025-10-14 15:34:57.530599503 +0000 UTC m=+1227.514647871" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.552483 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.562473 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.570576 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:34:57 crc kubenswrapper[4945]: E1014 15:34:57.571085 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48e8370d-1890-4a06-9566-5696dd1f926c" containerName="nova-scheduler-scheduler" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.571109 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="48e8370d-1890-4a06-9566-5696dd1f926c" containerName="nova-scheduler-scheduler" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.571373 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="48e8370d-1890-4a06-9566-5696dd1f926c" containerName="nova-scheduler-scheduler" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.572220 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.573944 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.601755 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.654438 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65e47e7-2cb8-492c-aef0-6bc069d1f643-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.654527 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8g84\" (UniqueName: \"kubernetes.io/projected/d65e47e7-2cb8-492c-aef0-6bc069d1f643-kube-api-access-t8g84\") pod \"nova-scheduler-0\" (UID: \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.654594 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65e47e7-2cb8-492c-aef0-6bc069d1f643-config-data\") pod \"nova-scheduler-0\" (UID: \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.756196 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65e47e7-2cb8-492c-aef0-6bc069d1f643-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.756259 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8g84\" (UniqueName: \"kubernetes.io/projected/d65e47e7-2cb8-492c-aef0-6bc069d1f643-kube-api-access-t8g84\") pod \"nova-scheduler-0\" (UID: \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.756327 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65e47e7-2cb8-492c-aef0-6bc069d1f643-config-data\") pod \"nova-scheduler-0\" (UID: \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.760368 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65e47e7-2cb8-492c-aef0-6bc069d1f643-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.766396 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65e47e7-2cb8-492c-aef0-6bc069d1f643-config-data\") pod \"nova-scheduler-0\" (UID: \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.772062 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8g84\" (UniqueName: \"kubernetes.io/projected/d65e47e7-2cb8-492c-aef0-6bc069d1f643-kube-api-access-t8g84\") pod \"nova-scheduler-0\" (UID: \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\") " pod="openstack/nova-scheduler-0" Oct 14 15:34:57 crc kubenswrapper[4945]: I1014 15:34:57.921204 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 15:34:58 crc kubenswrapper[4945]: I1014 15:34:58.394541 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:34:58 crc kubenswrapper[4945]: I1014 15:34:58.498736 4945 generic.go:334] "Generic (PLEG): container finished" podID="721e6f5e-9053-494a-80f7-127ad3ec46f8" containerID="3a272412c4f603eaf36663e4c94d6835d7c4c656a50f940b0ba53bed08eab148" exitCode=0 Oct 14 15:34:58 crc kubenswrapper[4945]: I1014 15:34:58.498812 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"721e6f5e-9053-494a-80f7-127ad3ec46f8","Type":"ContainerDied","Data":"3a272412c4f603eaf36663e4c94d6835d7c4c656a50f940b0ba53bed08eab148"} Oct 14 15:34:58 crc kubenswrapper[4945]: I1014 15:34:58.500496 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d65e47e7-2cb8-492c-aef0-6bc069d1f643","Type":"ContainerStarted","Data":"839af0f67b96485c0f0551546c8e9b2ae6f5f07080d420e0f95e6a2de858dd38"} Oct 14 15:34:58 crc kubenswrapper[4945]: I1014 15:34:58.781595 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48e8370d-1890-4a06-9566-5696dd1f926c" path="/var/lib/kubelet/pods/48e8370d-1890-4a06-9566-5696dd1f926c/volumes" Oct 14 15:34:58 crc kubenswrapper[4945]: I1014 15:34:58.976912 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.108078 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/721e6f5e-9053-494a-80f7-127ad3ec46f8-config-data\") pod \"721e6f5e-9053-494a-80f7-127ad3ec46f8\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.108207 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/721e6f5e-9053-494a-80f7-127ad3ec46f8-logs\") pod \"721e6f5e-9053-494a-80f7-127ad3ec46f8\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.108240 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721e6f5e-9053-494a-80f7-127ad3ec46f8-combined-ca-bundle\") pod \"721e6f5e-9053-494a-80f7-127ad3ec46f8\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.108322 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vs8gj\" (UniqueName: \"kubernetes.io/projected/721e6f5e-9053-494a-80f7-127ad3ec46f8-kube-api-access-vs8gj\") pod \"721e6f5e-9053-494a-80f7-127ad3ec46f8\" (UID: \"721e6f5e-9053-494a-80f7-127ad3ec46f8\") " Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.110106 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/721e6f5e-9053-494a-80f7-127ad3ec46f8-logs" (OuterVolumeSpecName: "logs") pod "721e6f5e-9053-494a-80f7-127ad3ec46f8" (UID: "721e6f5e-9053-494a-80f7-127ad3ec46f8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.122268 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/721e6f5e-9053-494a-80f7-127ad3ec46f8-kube-api-access-vs8gj" (OuterVolumeSpecName: "kube-api-access-vs8gj") pod "721e6f5e-9053-494a-80f7-127ad3ec46f8" (UID: "721e6f5e-9053-494a-80f7-127ad3ec46f8"). InnerVolumeSpecName "kube-api-access-vs8gj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.146285 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721e6f5e-9053-494a-80f7-127ad3ec46f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "721e6f5e-9053-494a-80f7-127ad3ec46f8" (UID: "721e6f5e-9053-494a-80f7-127ad3ec46f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.161111 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721e6f5e-9053-494a-80f7-127ad3ec46f8-config-data" (OuterVolumeSpecName: "config-data") pod "721e6f5e-9053-494a-80f7-127ad3ec46f8" (UID: "721e6f5e-9053-494a-80f7-127ad3ec46f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.211458 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/721e6f5e-9053-494a-80f7-127ad3ec46f8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.211702 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/721e6f5e-9053-494a-80f7-127ad3ec46f8-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.211824 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721e6f5e-9053-494a-80f7-127ad3ec46f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.212182 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vs8gj\" (UniqueName: \"kubernetes.io/projected/721e6f5e-9053-494a-80f7-127ad3ec46f8-kube-api-access-vs8gj\") on node \"crc\" DevicePath \"\"" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.517236 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.517240 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"721e6f5e-9053-494a-80f7-127ad3ec46f8","Type":"ContainerDied","Data":"8f180a3d12a48b27e33ef4d5951653422e0ddf2b449ad7e3f51e38383b77e088"} Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.517394 4945 scope.go:117] "RemoveContainer" containerID="3a272412c4f603eaf36663e4c94d6835d7c4c656a50f940b0ba53bed08eab148" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.521570 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d65e47e7-2cb8-492c-aef0-6bc069d1f643","Type":"ContainerStarted","Data":"6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4"} Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.547305 4945 scope.go:117] "RemoveContainer" containerID="1eeb489670fa6084db38300f253aac3144a9b047725e66c1648960fbb272d085" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.553798 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.5537493319999998 podStartE2EDuration="2.553749332s" podCreationTimestamp="2025-10-14 15:34:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:34:59.542271764 +0000 UTC m=+1229.526320162" watchObservedRunningTime="2025-10-14 15:34:59.553749332 +0000 UTC m=+1229.537797710" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.580742 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.589525 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.597928 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 15:34:59 crc kubenswrapper[4945]: E1014 15:34:59.598456 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="721e6f5e-9053-494a-80f7-127ad3ec46f8" containerName="nova-api-log" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.598478 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="721e6f5e-9053-494a-80f7-127ad3ec46f8" containerName="nova-api-log" Oct 14 15:34:59 crc kubenswrapper[4945]: E1014 15:34:59.598495 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="721e6f5e-9053-494a-80f7-127ad3ec46f8" containerName="nova-api-api" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.598504 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="721e6f5e-9053-494a-80f7-127ad3ec46f8" containerName="nova-api-api" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.598750 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="721e6f5e-9053-494a-80f7-127ad3ec46f8" containerName="nova-api-log" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.598776 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="721e6f5e-9053-494a-80f7-127ad3ec46f8" containerName="nova-api-api" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.600081 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.604178 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.651591 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.721356 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn5cr\" (UniqueName: \"kubernetes.io/projected/b83cf13e-faa9-46d6-a44e-560c56280f88-kube-api-access-xn5cr\") pod \"nova-api-0\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.721745 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b83cf13e-faa9-46d6-a44e-560c56280f88-logs\") pod \"nova-api-0\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.721796 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b83cf13e-faa9-46d6-a44e-560c56280f88-config-data\") pod \"nova-api-0\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.721826 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b83cf13e-faa9-46d6-a44e-560c56280f88-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.823549 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn5cr\" (UniqueName: \"kubernetes.io/projected/b83cf13e-faa9-46d6-a44e-560c56280f88-kube-api-access-xn5cr\") pod \"nova-api-0\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.823690 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b83cf13e-faa9-46d6-a44e-560c56280f88-logs\") pod \"nova-api-0\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.823711 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b83cf13e-faa9-46d6-a44e-560c56280f88-config-data\") pod \"nova-api-0\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.823725 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b83cf13e-faa9-46d6-a44e-560c56280f88-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.824407 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b83cf13e-faa9-46d6-a44e-560c56280f88-logs\") pod \"nova-api-0\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.828477 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b83cf13e-faa9-46d6-a44e-560c56280f88-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.830803 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b83cf13e-faa9-46d6-a44e-560c56280f88-config-data\") pod \"nova-api-0\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.840358 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn5cr\" (UniqueName: \"kubernetes.io/projected/b83cf13e-faa9-46d6-a44e-560c56280f88-kube-api-access-xn5cr\") pod \"nova-api-0\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " pod="openstack/nova-api-0" Oct 14 15:34:59 crc kubenswrapper[4945]: I1014 15:34:59.968967 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:35:00 crc kubenswrapper[4945]: I1014 15:35:00.436379 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:00 crc kubenswrapper[4945]: W1014 15:35:00.443273 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb83cf13e_faa9_46d6_a44e_560c56280f88.slice/crio-d2a30c3ea32e8635efa0899c2a6cb7d363948b4efbea4945f4879beb8316ce3f WatchSource:0}: Error finding container d2a30c3ea32e8635efa0899c2a6cb7d363948b4efbea4945f4879beb8316ce3f: Status 404 returned error can't find the container with id d2a30c3ea32e8635efa0899c2a6cb7d363948b4efbea4945f4879beb8316ce3f Oct 14 15:35:00 crc kubenswrapper[4945]: I1014 15:35:00.533010 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b83cf13e-faa9-46d6-a44e-560c56280f88","Type":"ContainerStarted","Data":"d2a30c3ea32e8635efa0899c2a6cb7d363948b4efbea4945f4879beb8316ce3f"} Oct 14 15:35:00 crc kubenswrapper[4945]: I1014 15:35:00.777308 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="721e6f5e-9053-494a-80f7-127ad3ec46f8" path="/var/lib/kubelet/pods/721e6f5e-9053-494a-80f7-127ad3ec46f8/volumes" Oct 14 15:35:00 crc kubenswrapper[4945]: I1014 15:35:00.927723 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 15:35:00 crc kubenswrapper[4945]: I1014 15:35:00.927832 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 15:35:01 crc kubenswrapper[4945]: I1014 15:35:01.543702 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b83cf13e-faa9-46d6-a44e-560c56280f88","Type":"ContainerStarted","Data":"59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca"} Oct 14 15:35:01 crc kubenswrapper[4945]: I1014 15:35:01.544094 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b83cf13e-faa9-46d6-a44e-560c56280f88","Type":"ContainerStarted","Data":"a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97"} Oct 14 15:35:01 crc kubenswrapper[4945]: I1014 15:35:01.573069 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.573045422 podStartE2EDuration="2.573045422s" podCreationTimestamp="2025-10-14 15:34:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:35:01.563632563 +0000 UTC m=+1231.547680961" watchObservedRunningTime="2025-10-14 15:35:01.573045422 +0000 UTC m=+1231.557093790" Oct 14 15:35:02 crc kubenswrapper[4945]: I1014 15:35:02.375658 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 15:35:02 crc kubenswrapper[4945]: I1014 15:35:02.921865 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 15:35:03 crc kubenswrapper[4945]: I1014 15:35:03.456358 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ironic-inspector-0" Oct 14 15:35:03 crc kubenswrapper[4945]: I1014 15:35:03.460915 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ironic-inspector-0" Oct 14 15:35:03 crc kubenswrapper[4945]: I1014 15:35:03.463020 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Oct 14 15:35:03 crc kubenswrapper[4945]: I1014 15:35:03.567557 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Oct 14 15:35:04 crc kubenswrapper[4945]: I1014 15:35:04.830595 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 14 15:35:05 crc kubenswrapper[4945]: I1014 15:35:05.927461 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 15:35:05 crc kubenswrapper[4945]: I1014 15:35:05.927812 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 15:35:06 crc kubenswrapper[4945]: I1014 15:35:06.440384 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 15:35:06 crc kubenswrapper[4945]: I1014 15:35:06.441260 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="42e0b677-3889-4be1-8fbb-8495a2239ade" containerName="kube-state-metrics" containerID="cri-o://350cba45b4f29b9a5afac152ad30712ee1e1881184f779bb06e575a83dd6e046" gracePeriod=30 Oct 14 15:35:06 crc kubenswrapper[4945]: I1014 15:35:06.592610 4945 generic.go:334] "Generic (PLEG): container finished" podID="42e0b677-3889-4be1-8fbb-8495a2239ade" containerID="350cba45b4f29b9a5afac152ad30712ee1e1881184f779bb06e575a83dd6e046" exitCode=2 Oct 14 15:35:06 crc kubenswrapper[4945]: I1014 15:35:06.592710 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"42e0b677-3889-4be1-8fbb-8495a2239ade","Type":"ContainerDied","Data":"350cba45b4f29b9a5afac152ad30712ee1e1881184f779bb06e575a83dd6e046"} Oct 14 15:35:06 crc kubenswrapper[4945]: I1014 15:35:06.944092 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5e25b244-f49c-43e0-9cdd-395f0906c15c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 15:35:06 crc kubenswrapper[4945]: I1014 15:35:06.944172 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5e25b244-f49c-43e0-9cdd-395f0906c15c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 15:35:06 crc kubenswrapper[4945]: I1014 15:35:06.985545 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.054430 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzhmq\" (UniqueName: \"kubernetes.io/projected/42e0b677-3889-4be1-8fbb-8495a2239ade-kube-api-access-wzhmq\") pod \"42e0b677-3889-4be1-8fbb-8495a2239ade\" (UID: \"42e0b677-3889-4be1-8fbb-8495a2239ade\") " Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.060779 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42e0b677-3889-4be1-8fbb-8495a2239ade-kube-api-access-wzhmq" (OuterVolumeSpecName: "kube-api-access-wzhmq") pod "42e0b677-3889-4be1-8fbb-8495a2239ade" (UID: "42e0b677-3889-4be1-8fbb-8495a2239ade"). InnerVolumeSpecName "kube-api-access-wzhmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.156667 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzhmq\" (UniqueName: \"kubernetes.io/projected/42e0b677-3889-4be1-8fbb-8495a2239ade-kube-api-access-wzhmq\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.604725 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"42e0b677-3889-4be1-8fbb-8495a2239ade","Type":"ContainerDied","Data":"06e855ff90a90dc86aacefe05ca369a96d6147838caff960e83c96a7a741d9c8"} Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.604780 4945 scope.go:117] "RemoveContainer" containerID="350cba45b4f29b9a5afac152ad30712ee1e1881184f779bb06e575a83dd6e046" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.604938 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.651564 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.653591 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.661050 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 15:35:07 crc kubenswrapper[4945]: E1014 15:35:07.661454 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e0b677-3889-4be1-8fbb-8495a2239ade" containerName="kube-state-metrics" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.661470 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e0b677-3889-4be1-8fbb-8495a2239ade" containerName="kube-state-metrics" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.661665 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="42e0b677-3889-4be1-8fbb-8495a2239ade" containerName="kube-state-metrics" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.662259 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.663889 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.664301 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.677571 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.766408 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41add807-ab1f-4d5f-999f-51626e4bb839-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"41add807-ab1f-4d5f-999f-51626e4bb839\") " pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.766460 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/41add807-ab1f-4d5f-999f-51626e4bb839-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"41add807-ab1f-4d5f-999f-51626e4bb839\") " pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.766504 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/41add807-ab1f-4d5f-999f-51626e4bb839-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"41add807-ab1f-4d5f-999f-51626e4bb839\") " pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.766602 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-944cw\" (UniqueName: \"kubernetes.io/projected/41add807-ab1f-4d5f-999f-51626e4bb839-kube-api-access-944cw\") pod \"kube-state-metrics-0\" (UID: \"41add807-ab1f-4d5f-999f-51626e4bb839\") " pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.867973 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-944cw\" (UniqueName: \"kubernetes.io/projected/41add807-ab1f-4d5f-999f-51626e4bb839-kube-api-access-944cw\") pod \"kube-state-metrics-0\" (UID: \"41add807-ab1f-4d5f-999f-51626e4bb839\") " pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.868371 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41add807-ab1f-4d5f-999f-51626e4bb839-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"41add807-ab1f-4d5f-999f-51626e4bb839\") " pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.868485 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/41add807-ab1f-4d5f-999f-51626e4bb839-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"41add807-ab1f-4d5f-999f-51626e4bb839\") " pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.868604 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/41add807-ab1f-4d5f-999f-51626e4bb839-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"41add807-ab1f-4d5f-999f-51626e4bb839\") " pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.873495 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/41add807-ab1f-4d5f-999f-51626e4bb839-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"41add807-ab1f-4d5f-999f-51626e4bb839\") " pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.873542 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/41add807-ab1f-4d5f-999f-51626e4bb839-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"41add807-ab1f-4d5f-999f-51626e4bb839\") " pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.873757 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41add807-ab1f-4d5f-999f-51626e4bb839-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"41add807-ab1f-4d5f-999f-51626e4bb839\") " pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.900505 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-944cw\" (UniqueName: \"kubernetes.io/projected/41add807-ab1f-4d5f-999f-51626e4bb839-kube-api-access-944cw\") pod \"kube-state-metrics-0\" (UID: \"41add807-ab1f-4d5f-999f-51626e4bb839\") " pod="openstack/kube-state-metrics-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.921990 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.956532 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 15:35:07 crc kubenswrapper[4945]: I1014 15:35:07.977227 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 15:35:08 crc kubenswrapper[4945]: I1014 15:35:08.439994 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 15:35:08 crc kubenswrapper[4945]: I1014 15:35:08.544828 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:08 crc kubenswrapper[4945]: I1014 15:35:08.545181 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="ceilometer-central-agent" containerID="cri-o://5f08462e4582c41377a16492373bfb746d8e2f683cec22a0936bc2127a6cd17f" gracePeriod=30 Oct 14 15:35:08 crc kubenswrapper[4945]: I1014 15:35:08.545726 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="proxy-httpd" containerID="cri-o://0844a95c766fb5cbba39086814548a0ac5a6aa9fc4f0da2ba2e65a1b860b9378" gracePeriod=30 Oct 14 15:35:08 crc kubenswrapper[4945]: I1014 15:35:08.545819 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="ceilometer-notification-agent" containerID="cri-o://318253a2d76677862b0788c33814f2b21b39907b7dc11ef78f43672fb9da3bb8" gracePeriod=30 Oct 14 15:35:08 crc kubenswrapper[4945]: I1014 15:35:08.545858 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="sg-core" containerID="cri-o://d53cbc3845fcf0ed2b50c66d9e6c5b082638890503f7a44548a519f13d3d076b" gracePeriod=30 Oct 14 15:35:08 crc kubenswrapper[4945]: I1014 15:35:08.615640 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"41add807-ab1f-4d5f-999f-51626e4bb839","Type":"ContainerStarted","Data":"25686972780be00b7aae14cc6dbd26427b039cf31956a66529c1c951f9beb94a"} Oct 14 15:35:08 crc kubenswrapper[4945]: I1014 15:35:08.643937 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 15:35:08 crc kubenswrapper[4945]: I1014 15:35:08.780015 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42e0b677-3889-4be1-8fbb-8495a2239ade" path="/var/lib/kubelet/pods/42e0b677-3889-4be1-8fbb-8495a2239ade/volumes" Oct 14 15:35:09 crc kubenswrapper[4945]: I1014 15:35:09.628424 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"41add807-ab1f-4d5f-999f-51626e4bb839","Type":"ContainerStarted","Data":"4b4f4782acf447e65a87281fbc31cdf51e014cfb4b55f502fd0abbb3d391dc0e"} Oct 14 15:35:09 crc kubenswrapper[4945]: I1014 15:35:09.629074 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 14 15:35:09 crc kubenswrapper[4945]: I1014 15:35:09.631414 4945 generic.go:334] "Generic (PLEG): container finished" podID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerID="0844a95c766fb5cbba39086814548a0ac5a6aa9fc4f0da2ba2e65a1b860b9378" exitCode=0 Oct 14 15:35:09 crc kubenswrapper[4945]: I1014 15:35:09.631449 4945 generic.go:334] "Generic (PLEG): container finished" podID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerID="d53cbc3845fcf0ed2b50c66d9e6c5b082638890503f7a44548a519f13d3d076b" exitCode=2 Oct 14 15:35:09 crc kubenswrapper[4945]: I1014 15:35:09.631459 4945 generic.go:334] "Generic (PLEG): container finished" podID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerID="5f08462e4582c41377a16492373bfb746d8e2f683cec22a0936bc2127a6cd17f" exitCode=0 Oct 14 15:35:09 crc kubenswrapper[4945]: I1014 15:35:09.631475 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fdab3dc-df19-4f56-a226-8afb55579e2e","Type":"ContainerDied","Data":"0844a95c766fb5cbba39086814548a0ac5a6aa9fc4f0da2ba2e65a1b860b9378"} Oct 14 15:35:09 crc kubenswrapper[4945]: I1014 15:35:09.631509 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fdab3dc-df19-4f56-a226-8afb55579e2e","Type":"ContainerDied","Data":"d53cbc3845fcf0ed2b50c66d9e6c5b082638890503f7a44548a519f13d3d076b"} Oct 14 15:35:09 crc kubenswrapper[4945]: I1014 15:35:09.631520 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fdab3dc-df19-4f56-a226-8afb55579e2e","Type":"ContainerDied","Data":"5f08462e4582c41377a16492373bfb746d8e2f683cec22a0936bc2127a6cd17f"} Oct 14 15:35:09 crc kubenswrapper[4945]: I1014 15:35:09.652949 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.304322901 podStartE2EDuration="2.652925085s" podCreationTimestamp="2025-10-14 15:35:07 +0000 UTC" firstStartedPulling="2025-10-14 15:35:08.450739957 +0000 UTC m=+1238.434788325" lastFinishedPulling="2025-10-14 15:35:08.799342141 +0000 UTC m=+1238.783390509" observedRunningTime="2025-10-14 15:35:09.643440304 +0000 UTC m=+1239.627488672" watchObservedRunningTime="2025-10-14 15:35:09.652925085 +0000 UTC m=+1239.636973463" Oct 14 15:35:09 crc kubenswrapper[4945]: I1014 15:35:09.969816 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 15:35:09 crc kubenswrapper[4945]: I1014 15:35:09.969895 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 15:35:11 crc kubenswrapper[4945]: I1014 15:35:11.054185 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b83cf13e-faa9-46d6-a44e-560c56280f88" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.203:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 15:35:11 crc kubenswrapper[4945]: I1014 15:35:11.054280 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b83cf13e-faa9-46d6-a44e-560c56280f88" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.203:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 15:35:11 crc kubenswrapper[4945]: I1014 15:35:11.651760 4945 generic.go:334] "Generic (PLEG): container finished" podID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerID="318253a2d76677862b0788c33814f2b21b39907b7dc11ef78f43672fb9da3bb8" exitCode=0 Oct 14 15:35:11 crc kubenswrapper[4945]: I1014 15:35:11.651805 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fdab3dc-df19-4f56-a226-8afb55579e2e","Type":"ContainerDied","Data":"318253a2d76677862b0788c33814f2b21b39907b7dc11ef78f43672fb9da3bb8"} Oct 14 15:35:11 crc kubenswrapper[4945]: I1014 15:35:11.806985 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="42e0b677-3889-4be1-8fbb-8495a2239ade" containerName="kube-state-metrics" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.089448 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.182861 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fdab3dc-df19-4f56-a226-8afb55579e2e-run-httpd\") pod \"9fdab3dc-df19-4f56-a226-8afb55579e2e\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.182966 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-scripts\") pod \"9fdab3dc-df19-4f56-a226-8afb55579e2e\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.183000 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp262\" (UniqueName: \"kubernetes.io/projected/9fdab3dc-df19-4f56-a226-8afb55579e2e-kube-api-access-xp262\") pod \"9fdab3dc-df19-4f56-a226-8afb55579e2e\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.183107 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-config-data\") pod \"9fdab3dc-df19-4f56-a226-8afb55579e2e\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.183145 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fdab3dc-df19-4f56-a226-8afb55579e2e-log-httpd\") pod \"9fdab3dc-df19-4f56-a226-8afb55579e2e\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.183214 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-combined-ca-bundle\") pod \"9fdab3dc-df19-4f56-a226-8afb55579e2e\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.183518 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fdab3dc-df19-4f56-a226-8afb55579e2e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9fdab3dc-df19-4f56-a226-8afb55579e2e" (UID: "9fdab3dc-df19-4f56-a226-8afb55579e2e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.184049 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fdab3dc-df19-4f56-a226-8afb55579e2e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9fdab3dc-df19-4f56-a226-8afb55579e2e" (UID: "9fdab3dc-df19-4f56-a226-8afb55579e2e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.184288 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-sg-core-conf-yaml\") pod \"9fdab3dc-df19-4f56-a226-8afb55579e2e\" (UID: \"9fdab3dc-df19-4f56-a226-8afb55579e2e\") " Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.186112 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fdab3dc-df19-4f56-a226-8afb55579e2e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.186134 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fdab3dc-df19-4f56-a226-8afb55579e2e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.201826 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-scripts" (OuterVolumeSpecName: "scripts") pod "9fdab3dc-df19-4f56-a226-8afb55579e2e" (UID: "9fdab3dc-df19-4f56-a226-8afb55579e2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.205246 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fdab3dc-df19-4f56-a226-8afb55579e2e-kube-api-access-xp262" (OuterVolumeSpecName: "kube-api-access-xp262") pod "9fdab3dc-df19-4f56-a226-8afb55579e2e" (UID: "9fdab3dc-df19-4f56-a226-8afb55579e2e"). InnerVolumeSpecName "kube-api-access-xp262". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.217066 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9fdab3dc-df19-4f56-a226-8afb55579e2e" (UID: "9fdab3dc-df19-4f56-a226-8afb55579e2e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.295798 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.295833 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xp262\" (UniqueName: \"kubernetes.io/projected/9fdab3dc-df19-4f56-a226-8afb55579e2e-kube-api-access-xp262\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.295849 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.405775 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fdab3dc-df19-4f56-a226-8afb55579e2e" (UID: "9fdab3dc-df19-4f56-a226-8afb55579e2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.483525 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-config-data" (OuterVolumeSpecName: "config-data") pod "9fdab3dc-df19-4f56-a226-8afb55579e2e" (UID: "9fdab3dc-df19-4f56-a226-8afb55579e2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.500415 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.500446 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdab3dc-df19-4f56-a226-8afb55579e2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.679414 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fdab3dc-df19-4f56-a226-8afb55579e2e","Type":"ContainerDied","Data":"51f7433478f6b9d700a1ecb35580876b1c33e0d42b1c7e7e89e52d465dc62bba"} Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.679475 4945 scope.go:117] "RemoveContainer" containerID="0844a95c766fb5cbba39086814548a0ac5a6aa9fc4f0da2ba2e65a1b860b9378" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.679473 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.705572 4945 scope.go:117] "RemoveContainer" containerID="d53cbc3845fcf0ed2b50c66d9e6c5b082638890503f7a44548a519f13d3d076b" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.727262 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.730917 4945 scope.go:117] "RemoveContainer" containerID="318253a2d76677862b0788c33814f2b21b39907b7dc11ef78f43672fb9da3bb8" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.741781 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.755316 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:12 crc kubenswrapper[4945]: E1014 15:35:12.755734 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="ceilometer-notification-agent" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.755755 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="ceilometer-notification-agent" Oct 14 15:35:12 crc kubenswrapper[4945]: E1014 15:35:12.755770 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="sg-core" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.755777 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="sg-core" Oct 14 15:35:12 crc kubenswrapper[4945]: E1014 15:35:12.755787 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="proxy-httpd" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.755793 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="proxy-httpd" Oct 14 15:35:12 crc kubenswrapper[4945]: E1014 15:35:12.755814 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="ceilometer-central-agent" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.755820 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="ceilometer-central-agent" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.756014 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="ceilometer-notification-agent" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.756026 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="ceilometer-central-agent" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.756037 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="sg-core" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.756043 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" containerName="proxy-httpd" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.757636 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.762513 4945 scope.go:117] "RemoveContainer" containerID="5f08462e4582c41377a16492373bfb746d8e2f683cec22a0936bc2127a6cd17f" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.762651 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.762766 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.762855 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.775767 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fdab3dc-df19-4f56-a226-8afb55579e2e" path="/var/lib/kubelet/pods/9fdab3dc-df19-4f56-a226-8afb55579e2e/volumes" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.776991 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.906708 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.907000 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d9c526-635a-4587-b099-43234aa76dfd-log-httpd\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.907113 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.907389 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbtgn\" (UniqueName: \"kubernetes.io/projected/70d9c526-635a-4587-b099-43234aa76dfd-kube-api-access-hbtgn\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.907993 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d9c526-635a-4587-b099-43234aa76dfd-run-httpd\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.908072 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-scripts\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.908104 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-config-data\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:12 crc kubenswrapper[4945]: I1014 15:35:12.908135 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.010401 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbtgn\" (UniqueName: \"kubernetes.io/projected/70d9c526-635a-4587-b099-43234aa76dfd-kube-api-access-hbtgn\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.010795 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d9c526-635a-4587-b099-43234aa76dfd-run-httpd\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.010936 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-scripts\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.011322 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d9c526-635a-4587-b099-43234aa76dfd-run-httpd\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.011365 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-config-data\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.011421 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.011481 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.011514 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d9c526-635a-4587-b099-43234aa76dfd-log-httpd\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.011536 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.012177 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d9c526-635a-4587-b099-43234aa76dfd-log-httpd\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.017679 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-scripts\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.018240 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.019481 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.021380 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-config-data\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.022211 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.035749 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbtgn\" (UniqueName: \"kubernetes.io/projected/70d9c526-635a-4587-b099-43234aa76dfd-kube-api-access-hbtgn\") pod \"ceilometer-0\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.081535 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.596363 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:13 crc kubenswrapper[4945]: I1014 15:35:13.689474 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70d9c526-635a-4587-b099-43234aa76dfd","Type":"ContainerStarted","Data":"77b2b2525dd782978809abe9bbfb1827aae05388de08e7422a3660be97ee9bc4"} Oct 14 15:35:15 crc kubenswrapper[4945]: I1014 15:35:15.709430 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70d9c526-635a-4587-b099-43234aa76dfd","Type":"ContainerStarted","Data":"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8"} Oct 14 15:35:15 crc kubenswrapper[4945]: I1014 15:35:15.709895 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70d9c526-635a-4587-b099-43234aa76dfd","Type":"ContainerStarted","Data":"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05"} Oct 14 15:35:15 crc kubenswrapper[4945]: I1014 15:35:15.933065 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 15:35:15 crc kubenswrapper[4945]: I1014 15:35:15.933715 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 15:35:15 crc kubenswrapper[4945]: I1014 15:35:15.938892 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 15:35:15 crc kubenswrapper[4945]: I1014 15:35:15.941416 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 15:35:16 crc kubenswrapper[4945]: I1014 15:35:16.720137 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70d9c526-635a-4587-b099-43234aa76dfd","Type":"ContainerStarted","Data":"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4"} Oct 14 15:35:16 crc kubenswrapper[4945]: I1014 15:35:16.797075 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:35:16 crc kubenswrapper[4945]: I1014 15:35:16.797371 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:35:17 crc kubenswrapper[4945]: I1014 15:35:17.987513 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 14 15:35:19 crc kubenswrapper[4945]: I1014 15:35:19.750015 4945 generic.go:334] "Generic (PLEG): container finished" podID="314ac186-b037-4e81-953c-78288edadcd4" containerID="aadf692b7b29f8ed5a3f733bb5119e548fb07857bdc3e69d009b43a143e29f62" exitCode=137 Oct 14 15:35:19 crc kubenswrapper[4945]: I1014 15:35:19.750121 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"314ac186-b037-4e81-953c-78288edadcd4","Type":"ContainerDied","Data":"aadf692b7b29f8ed5a3f733bb5119e548fb07857bdc3e69d009b43a143e29f62"} Oct 14 15:35:19 crc kubenswrapper[4945]: I1014 15:35:19.983482 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 15:35:19 crc kubenswrapper[4945]: I1014 15:35:19.984365 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 15:35:19 crc kubenswrapper[4945]: I1014 15:35:19.990235 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 15:35:19 crc kubenswrapper[4945]: I1014 15:35:19.991059 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.213727 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.353035 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n8bq\" (UniqueName: \"kubernetes.io/projected/314ac186-b037-4e81-953c-78288edadcd4-kube-api-access-9n8bq\") pod \"314ac186-b037-4e81-953c-78288edadcd4\" (UID: \"314ac186-b037-4e81-953c-78288edadcd4\") " Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.353095 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/314ac186-b037-4e81-953c-78288edadcd4-combined-ca-bundle\") pod \"314ac186-b037-4e81-953c-78288edadcd4\" (UID: \"314ac186-b037-4e81-953c-78288edadcd4\") " Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.353330 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/314ac186-b037-4e81-953c-78288edadcd4-config-data\") pod \"314ac186-b037-4e81-953c-78288edadcd4\" (UID: \"314ac186-b037-4e81-953c-78288edadcd4\") " Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.360019 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/314ac186-b037-4e81-953c-78288edadcd4-kube-api-access-9n8bq" (OuterVolumeSpecName: "kube-api-access-9n8bq") pod "314ac186-b037-4e81-953c-78288edadcd4" (UID: "314ac186-b037-4e81-953c-78288edadcd4"). InnerVolumeSpecName "kube-api-access-9n8bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.379539 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/314ac186-b037-4e81-953c-78288edadcd4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "314ac186-b037-4e81-953c-78288edadcd4" (UID: "314ac186-b037-4e81-953c-78288edadcd4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.381219 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/314ac186-b037-4e81-953c-78288edadcd4-config-data" (OuterVolumeSpecName: "config-data") pod "314ac186-b037-4e81-953c-78288edadcd4" (UID: "314ac186-b037-4e81-953c-78288edadcd4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.454992 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/314ac186-b037-4e81-953c-78288edadcd4-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.455025 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n8bq\" (UniqueName: \"kubernetes.io/projected/314ac186-b037-4e81-953c-78288edadcd4-kube-api-access-9n8bq\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.455047 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/314ac186-b037-4e81-953c-78288edadcd4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.772457 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.774237 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"314ac186-b037-4e81-953c-78288edadcd4","Type":"ContainerDied","Data":"c6650db9e256dc0912d1a7061aa7d6dd6fdc26d7e8ebba6a07d891f689fc31cf"} Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.774284 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70d9c526-635a-4587-b099-43234aa76dfd","Type":"ContainerStarted","Data":"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf"} Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.774304 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.774347 4945 scope.go:117] "RemoveContainer" containerID="aadf692b7b29f8ed5a3f733bb5119e548fb07857bdc3e69d009b43a143e29f62" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.795313 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.820043 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.6326830279999998 podStartE2EDuration="8.820017481s" podCreationTimestamp="2025-10-14 15:35:12 +0000 UTC" firstStartedPulling="2025-10-14 15:35:13.603116399 +0000 UTC m=+1243.587164767" lastFinishedPulling="2025-10-14 15:35:19.790450832 +0000 UTC m=+1249.774499220" observedRunningTime="2025-10-14 15:35:20.817246172 +0000 UTC m=+1250.801294550" watchObservedRunningTime="2025-10-14 15:35:20.820017481 +0000 UTC m=+1250.804065849" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.845017 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.871993 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.923685 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 15:35:20 crc kubenswrapper[4945]: E1014 15:35:20.924425 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314ac186-b037-4e81-953c-78288edadcd4" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.924454 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="314ac186-b037-4e81-953c-78288edadcd4" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.924891 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="314ac186-b037-4e81-953c-78288edadcd4" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.925776 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.930262 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.930504 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 14 15:35:20 crc kubenswrapper[4945]: I1014 15:35:20.943272 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.013956 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.115322 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9af8e27a-69da-4094-925f-0d018b30d573-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.115379 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9af8e27a-69da-4094-925f-0d018b30d573-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.115430 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9af8e27a-69da-4094-925f-0d018b30d573-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.115516 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af8e27a-69da-4094-925f-0d018b30d573-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.115537 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs7c9\" (UniqueName: \"kubernetes.io/projected/9af8e27a-69da-4094-925f-0d018b30d573-kube-api-access-rs7c9\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.187045 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-hhtr4"] Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.189057 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.211387 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-hhtr4"] Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.218104 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9af8e27a-69da-4094-925f-0d018b30d573-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.218230 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af8e27a-69da-4094-925f-0d018b30d573-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.218251 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs7c9\" (UniqueName: \"kubernetes.io/projected/9af8e27a-69da-4094-925f-0d018b30d573-kube-api-access-rs7c9\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.218379 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9af8e27a-69da-4094-925f-0d018b30d573-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.218398 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9af8e27a-69da-4094-925f-0d018b30d573-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.230464 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af8e27a-69da-4094-925f-0d018b30d573-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.231284 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9af8e27a-69da-4094-925f-0d018b30d573-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.232405 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9af8e27a-69da-4094-925f-0d018b30d573-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.244038 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9af8e27a-69da-4094-925f-0d018b30d573-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.250571 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs7c9\" (UniqueName: \"kubernetes.io/projected/9af8e27a-69da-4094-925f-0d018b30d573-kube-api-access-rs7c9\") pod \"nova-cell1-novncproxy-0\" (UID: \"9af8e27a-69da-4094-925f-0d018b30d573\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.304217 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.321948 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.322002 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drcsq\" (UniqueName: \"kubernetes.io/projected/9b5e61fe-2a22-4128-90c3-569e2e83a972-kube-api-access-drcsq\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.322030 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.322205 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-config\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.322379 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.322685 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.427097 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.427150 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.427184 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drcsq\" (UniqueName: \"kubernetes.io/projected/9b5e61fe-2a22-4128-90c3-569e2e83a972-kube-api-access-drcsq\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.427200 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.427251 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-config\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.427315 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.428200 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.428749 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.429357 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.431493 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-config\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.431763 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b5e61fe-2a22-4128-90c3-569e2e83a972-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.448788 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drcsq\" (UniqueName: \"kubernetes.io/projected/9b5e61fe-2a22-4128-90c3-569e2e83a972-kube-api-access-drcsq\") pod \"dnsmasq-dns-5c7b6c5df9-hhtr4\" (UID: \"9b5e61fe-2a22-4128-90c3-569e2e83a972\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.517482 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.780967 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 15:35:21 crc kubenswrapper[4945]: I1014 15:35:21.913014 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 15:35:22 crc kubenswrapper[4945]: W1014 15:35:22.054156 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b5e61fe_2a22_4128_90c3_569e2e83a972.slice/crio-af72e61bbeb6d2d9ef1aa66cfc5194facd37c4a7312d1fe90baacdb303a14a85 WatchSource:0}: Error finding container af72e61bbeb6d2d9ef1aa66cfc5194facd37c4a7312d1fe90baacdb303a14a85: Status 404 returned error can't find the container with id af72e61bbeb6d2d9ef1aa66cfc5194facd37c4a7312d1fe90baacdb303a14a85 Oct 14 15:35:22 crc kubenswrapper[4945]: I1014 15:35:22.054322 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-hhtr4"] Oct 14 15:35:22 crc kubenswrapper[4945]: I1014 15:35:22.772639 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="314ac186-b037-4e81-953c-78288edadcd4" path="/var/lib/kubelet/pods/314ac186-b037-4e81-953c-78288edadcd4/volumes" Oct 14 15:35:22 crc kubenswrapper[4945]: I1014 15:35:22.790411 4945 generic.go:334] "Generic (PLEG): container finished" podID="9b5e61fe-2a22-4128-90c3-569e2e83a972" containerID="017f7e8c60de4cc632fff6d3e5190f3a4b0563a881a65b5d5e66666398f4fdbd" exitCode=0 Oct 14 15:35:22 crc kubenswrapper[4945]: I1014 15:35:22.790477 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" event={"ID":"9b5e61fe-2a22-4128-90c3-569e2e83a972","Type":"ContainerDied","Data":"017f7e8c60de4cc632fff6d3e5190f3a4b0563a881a65b5d5e66666398f4fdbd"} Oct 14 15:35:22 crc kubenswrapper[4945]: I1014 15:35:22.790503 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" event={"ID":"9b5e61fe-2a22-4128-90c3-569e2e83a972","Type":"ContainerStarted","Data":"af72e61bbeb6d2d9ef1aa66cfc5194facd37c4a7312d1fe90baacdb303a14a85"} Oct 14 15:35:22 crc kubenswrapper[4945]: I1014 15:35:22.795492 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9af8e27a-69da-4094-925f-0d018b30d573","Type":"ContainerStarted","Data":"12f987b5439938b719e55257a8c848f041209e8dd5d8b34c595a1868d32e6db9"} Oct 14 15:35:22 crc kubenswrapper[4945]: I1014 15:35:22.795542 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9af8e27a-69da-4094-925f-0d018b30d573","Type":"ContainerStarted","Data":"5dd6a757c58034e0e58587a72f409b9da1ce9f194ca5e43fd69e7f0861afeb1e"} Oct 14 15:35:22 crc kubenswrapper[4945]: I1014 15:35:22.861304 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.861283927 podStartE2EDuration="2.861283927s" podCreationTimestamp="2025-10-14 15:35:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:35:22.847503413 +0000 UTC m=+1252.831551781" watchObservedRunningTime="2025-10-14 15:35:22.861283927 +0000 UTC m=+1252.845332295" Oct 14 15:35:23 crc kubenswrapper[4945]: I1014 15:35:23.805181 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" event={"ID":"9b5e61fe-2a22-4128-90c3-569e2e83a972","Type":"ContainerStarted","Data":"2cf533d7bf290c01182391d64458e3ce761ff4c3bc6c0ecc9e17d79b4884e7de"} Oct 14 15:35:23 crc kubenswrapper[4945]: I1014 15:35:23.832373 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" podStartSLOduration=2.832351995 podStartE2EDuration="2.832351995s" podCreationTimestamp="2025-10-14 15:35:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:35:23.823249135 +0000 UTC m=+1253.807297503" watchObservedRunningTime="2025-10-14 15:35:23.832351995 +0000 UTC m=+1253.816400383" Oct 14 15:35:23 crc kubenswrapper[4945]: I1014 15:35:23.936774 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:23 crc kubenswrapper[4945]: I1014 15:35:23.937034 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b83cf13e-faa9-46d6-a44e-560c56280f88" containerName="nova-api-log" containerID="cri-o://a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97" gracePeriod=30 Oct 14 15:35:23 crc kubenswrapper[4945]: I1014 15:35:23.937119 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b83cf13e-faa9-46d6-a44e-560c56280f88" containerName="nova-api-api" containerID="cri-o://59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca" gracePeriod=30 Oct 14 15:35:24 crc kubenswrapper[4945]: I1014 15:35:24.827786 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:24 crc kubenswrapper[4945]: I1014 15:35:24.829077 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="ceilometer-central-agent" containerID="cri-o://28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05" gracePeriod=30 Oct 14 15:35:24 crc kubenswrapper[4945]: I1014 15:35:24.829310 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="proxy-httpd" containerID="cri-o://69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf" gracePeriod=30 Oct 14 15:35:24 crc kubenswrapper[4945]: I1014 15:35:24.829377 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="sg-core" containerID="cri-o://5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4" gracePeriod=30 Oct 14 15:35:24 crc kubenswrapper[4945]: I1014 15:35:24.829419 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="ceilometer-notification-agent" containerID="cri-o://456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8" gracePeriod=30 Oct 14 15:35:24 crc kubenswrapper[4945]: I1014 15:35:24.830397 4945 generic.go:334] "Generic (PLEG): container finished" podID="b83cf13e-faa9-46d6-a44e-560c56280f88" containerID="a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97" exitCode=143 Oct 14 15:35:24 crc kubenswrapper[4945]: I1014 15:35:24.830437 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b83cf13e-faa9-46d6-a44e-560c56280f88","Type":"ContainerDied","Data":"a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97"} Oct 14 15:35:24 crc kubenswrapper[4945]: I1014 15:35:24.830608 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.764663 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.838561 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbtgn\" (UniqueName: \"kubernetes.io/projected/70d9c526-635a-4587-b099-43234aa76dfd-kube-api-access-hbtgn\") pod \"70d9c526-635a-4587-b099-43234aa76dfd\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.838650 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d9c526-635a-4587-b099-43234aa76dfd-log-httpd\") pod \"70d9c526-635a-4587-b099-43234aa76dfd\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.838765 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-combined-ca-bundle\") pod \"70d9c526-635a-4587-b099-43234aa76dfd\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.838811 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-scripts\") pod \"70d9c526-635a-4587-b099-43234aa76dfd\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.838914 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-sg-core-conf-yaml\") pod \"70d9c526-635a-4587-b099-43234aa76dfd\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.838951 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-config-data\") pod \"70d9c526-635a-4587-b099-43234aa76dfd\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.838981 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-ceilometer-tls-certs\") pod \"70d9c526-635a-4587-b099-43234aa76dfd\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.839082 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d9c526-635a-4587-b099-43234aa76dfd-run-httpd\") pod \"70d9c526-635a-4587-b099-43234aa76dfd\" (UID: \"70d9c526-635a-4587-b099-43234aa76dfd\") " Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.839368 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70d9c526-635a-4587-b099-43234aa76dfd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "70d9c526-635a-4587-b099-43234aa76dfd" (UID: "70d9c526-635a-4587-b099-43234aa76dfd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.839921 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d9c526-635a-4587-b099-43234aa76dfd-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.846996 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70d9c526-635a-4587-b099-43234aa76dfd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "70d9c526-635a-4587-b099-43234aa76dfd" (UID: "70d9c526-635a-4587-b099-43234aa76dfd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.856656 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70d9c526-635a-4587-b099-43234aa76dfd-kube-api-access-hbtgn" (OuterVolumeSpecName: "kube-api-access-hbtgn") pod "70d9c526-635a-4587-b099-43234aa76dfd" (UID: "70d9c526-635a-4587-b099-43234aa76dfd"). InnerVolumeSpecName "kube-api-access-hbtgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.860581 4945 generic.go:334] "Generic (PLEG): container finished" podID="70d9c526-635a-4587-b099-43234aa76dfd" containerID="69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf" exitCode=0 Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.860617 4945 generic.go:334] "Generic (PLEG): container finished" podID="70d9c526-635a-4587-b099-43234aa76dfd" containerID="5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4" exitCode=2 Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.860629 4945 generic.go:334] "Generic (PLEG): container finished" podID="70d9c526-635a-4587-b099-43234aa76dfd" containerID="456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8" exitCode=0 Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.860638 4945 generic.go:334] "Generic (PLEG): container finished" podID="70d9c526-635a-4587-b099-43234aa76dfd" containerID="28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05" exitCode=0 Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.861841 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.862408 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70d9c526-635a-4587-b099-43234aa76dfd","Type":"ContainerDied","Data":"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf"} Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.862448 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70d9c526-635a-4587-b099-43234aa76dfd","Type":"ContainerDied","Data":"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4"} Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.862465 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70d9c526-635a-4587-b099-43234aa76dfd","Type":"ContainerDied","Data":"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8"} Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.862478 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70d9c526-635a-4587-b099-43234aa76dfd","Type":"ContainerDied","Data":"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05"} Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.862489 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70d9c526-635a-4587-b099-43234aa76dfd","Type":"ContainerDied","Data":"77b2b2525dd782978809abe9bbfb1827aae05388de08e7422a3660be97ee9bc4"} Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.862508 4945 scope.go:117] "RemoveContainer" containerID="69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.867286 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-scripts" (OuterVolumeSpecName: "scripts") pod "70d9c526-635a-4587-b099-43234aa76dfd" (UID: "70d9c526-635a-4587-b099-43234aa76dfd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.887066 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "70d9c526-635a-4587-b099-43234aa76dfd" (UID: "70d9c526-635a-4587-b099-43234aa76dfd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.917000 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "70d9c526-635a-4587-b099-43234aa76dfd" (UID: "70d9c526-635a-4587-b099-43234aa76dfd"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.938478 4945 scope.go:117] "RemoveContainer" containerID="5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.941682 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.941715 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.941731 4945 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.941786 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70d9c526-635a-4587-b099-43234aa76dfd-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.941903 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbtgn\" (UniqueName: \"kubernetes.io/projected/70d9c526-635a-4587-b099-43234aa76dfd-kube-api-access-hbtgn\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.955588 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70d9c526-635a-4587-b099-43234aa76dfd" (UID: "70d9c526-635a-4587-b099-43234aa76dfd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.965256 4945 scope.go:117] "RemoveContainer" containerID="456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.986950 4945 scope.go:117] "RemoveContainer" containerID="28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05" Oct 14 15:35:25 crc kubenswrapper[4945]: I1014 15:35:25.995438 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-config-data" (OuterVolumeSpecName: "config-data") pod "70d9c526-635a-4587-b099-43234aa76dfd" (UID: "70d9c526-635a-4587-b099-43234aa76dfd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.010009 4945 scope.go:117] "RemoveContainer" containerID="69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf" Oct 14 15:35:26 crc kubenswrapper[4945]: E1014 15:35:26.010469 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf\": container with ID starting with 69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf not found: ID does not exist" containerID="69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.010500 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf"} err="failed to get container status \"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf\": rpc error: code = NotFound desc = could not find container \"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf\": container with ID starting with 69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.010523 4945 scope.go:117] "RemoveContainer" containerID="5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4" Oct 14 15:35:26 crc kubenswrapper[4945]: E1014 15:35:26.011013 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4\": container with ID starting with 5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4 not found: ID does not exist" containerID="5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.011041 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4"} err="failed to get container status \"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4\": rpc error: code = NotFound desc = could not find container \"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4\": container with ID starting with 5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4 not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.011054 4945 scope.go:117] "RemoveContainer" containerID="456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8" Oct 14 15:35:26 crc kubenswrapper[4945]: E1014 15:35:26.011601 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8\": container with ID starting with 456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8 not found: ID does not exist" containerID="456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.011623 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8"} err="failed to get container status \"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8\": rpc error: code = NotFound desc = could not find container \"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8\": container with ID starting with 456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8 not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.011638 4945 scope.go:117] "RemoveContainer" containerID="28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05" Oct 14 15:35:26 crc kubenswrapper[4945]: E1014 15:35:26.011927 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05\": container with ID starting with 28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05 not found: ID does not exist" containerID="28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.011951 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05"} err="failed to get container status \"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05\": rpc error: code = NotFound desc = could not find container \"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05\": container with ID starting with 28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05 not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.011965 4945 scope.go:117] "RemoveContainer" containerID="69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.012556 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf"} err="failed to get container status \"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf\": rpc error: code = NotFound desc = could not find container \"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf\": container with ID starting with 69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.012577 4945 scope.go:117] "RemoveContainer" containerID="5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.012948 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4"} err="failed to get container status \"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4\": rpc error: code = NotFound desc = could not find container \"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4\": container with ID starting with 5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4 not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.012998 4945 scope.go:117] "RemoveContainer" containerID="456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.013288 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8"} err="failed to get container status \"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8\": rpc error: code = NotFound desc = could not find container \"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8\": container with ID starting with 456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8 not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.013324 4945 scope.go:117] "RemoveContainer" containerID="28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.013604 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05"} err="failed to get container status \"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05\": rpc error: code = NotFound desc = could not find container \"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05\": container with ID starting with 28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05 not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.013635 4945 scope.go:117] "RemoveContainer" containerID="69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.013908 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf"} err="failed to get container status \"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf\": rpc error: code = NotFound desc = could not find container \"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf\": container with ID starting with 69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.013938 4945 scope.go:117] "RemoveContainer" containerID="5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.014145 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4"} err="failed to get container status \"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4\": rpc error: code = NotFound desc = could not find container \"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4\": container with ID starting with 5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4 not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.014166 4945 scope.go:117] "RemoveContainer" containerID="456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.014403 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8"} err="failed to get container status \"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8\": rpc error: code = NotFound desc = could not find container \"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8\": container with ID starting with 456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8 not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.014430 4945 scope.go:117] "RemoveContainer" containerID="28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.014643 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05"} err="failed to get container status \"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05\": rpc error: code = NotFound desc = could not find container \"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05\": container with ID starting with 28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05 not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.014662 4945 scope.go:117] "RemoveContainer" containerID="69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.015469 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf"} err="failed to get container status \"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf\": rpc error: code = NotFound desc = could not find container \"69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf\": container with ID starting with 69681a4df9e5b19d153a2ab0f8657f2940d856d6166eb61801515781171889bf not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.015495 4945 scope.go:117] "RemoveContainer" containerID="5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.015743 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4"} err="failed to get container status \"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4\": rpc error: code = NotFound desc = could not find container \"5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4\": container with ID starting with 5d820007b56d43d701336a7d5def44f0a80833293a5d055b2ebc3db7cbb1aca4 not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.015766 4945 scope.go:117] "RemoveContainer" containerID="456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.016014 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8"} err="failed to get container status \"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8\": rpc error: code = NotFound desc = could not find container \"456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8\": container with ID starting with 456d8bc5d6f4ff8b7ab972f07ba256918aba5a9b65097548783be3f4f7f922b8 not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.016042 4945 scope.go:117] "RemoveContainer" containerID="28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.016282 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05"} err="failed to get container status \"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05\": rpc error: code = NotFound desc = could not find container \"28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05\": container with ID starting with 28926709a842fc249980d6e3f96a602e62baaf06fbbc1f039ee02aa3d1097a05 not found: ID does not exist" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.043490 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.043526 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d9c526-635a-4587-b099-43234aa76dfd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.201607 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.215012 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.226669 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:26 crc kubenswrapper[4945]: E1014 15:35:26.227116 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="ceilometer-central-agent" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.227141 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="ceilometer-central-agent" Oct 14 15:35:26 crc kubenswrapper[4945]: E1014 15:35:26.227166 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="proxy-httpd" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.227175 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="proxy-httpd" Oct 14 15:35:26 crc kubenswrapper[4945]: E1014 15:35:26.227205 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="ceilometer-notification-agent" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.227213 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="ceilometer-notification-agent" Oct 14 15:35:26 crc kubenswrapper[4945]: E1014 15:35:26.227228 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="sg-core" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.227235 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="sg-core" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.227471 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="proxy-httpd" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.227489 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="ceilometer-notification-agent" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.227515 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="sg-core" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.227534 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="70d9c526-635a-4587-b099-43234aa76dfd" containerName="ceilometer-central-agent" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.230149 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.232606 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.232635 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.234964 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.240486 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.246612 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae81785-b556-4bd4-8a97-c0c5d9a211a6-run-httpd\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.246668 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-config-data\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.246710 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.246740 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.246757 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae81785-b556-4bd4-8a97-c0c5d9a211a6-log-httpd\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.246785 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.246799 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5fsx\" (UniqueName: \"kubernetes.io/projected/aae81785-b556-4bd4-8a97-c0c5d9a211a6-kube-api-access-b5fsx\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.246831 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-scripts\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.305228 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.348384 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae81785-b556-4bd4-8a97-c0c5d9a211a6-run-httpd\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.348458 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-config-data\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.348549 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.348598 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.348626 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae81785-b556-4bd4-8a97-c0c5d9a211a6-log-httpd\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.348656 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.348688 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5fsx\" (UniqueName: \"kubernetes.io/projected/aae81785-b556-4bd4-8a97-c0c5d9a211a6-kube-api-access-b5fsx\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.348833 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae81785-b556-4bd4-8a97-c0c5d9a211a6-run-httpd\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.349321 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae81785-b556-4bd4-8a97-c0c5d9a211a6-log-httpd\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.349819 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-scripts\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.353024 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.353764 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.354932 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.362158 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-scripts\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.363029 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-config-data\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.364745 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5fsx\" (UniqueName: \"kubernetes.io/projected/aae81785-b556-4bd4-8a97-c0c5d9a211a6-kube-api-access-b5fsx\") pod \"ceilometer-0\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.613450 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.735952 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:26 crc kubenswrapper[4945]: I1014 15:35:26.775316 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70d9c526-635a-4587-b099-43234aa76dfd" path="/var/lib/kubelet/pods/70d9c526-635a-4587-b099-43234aa76dfd/volumes" Oct 14 15:35:27 crc kubenswrapper[4945]: W1014 15:35:27.101887 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaae81785_b556_4bd4_8a97_c0c5d9a211a6.slice/crio-cc33750062c40ad3dcd32375b0d9aa0c6c544e8805e5e243c92807b274f7de98 WatchSource:0}: Error finding container cc33750062c40ad3dcd32375b0d9aa0c6c544e8805e5e243c92807b274f7de98: Status 404 returned error can't find the container with id cc33750062c40ad3dcd32375b0d9aa0c6c544e8805e5e243c92807b274f7de98 Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.106536 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.106628 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.810266 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.932249 4945 generic.go:334] "Generic (PLEG): container finished" podID="b83cf13e-faa9-46d6-a44e-560c56280f88" containerID="59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca" exitCode=0 Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.932308 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b83cf13e-faa9-46d6-a44e-560c56280f88","Type":"ContainerDied","Data":"59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca"} Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.932333 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b83cf13e-faa9-46d6-a44e-560c56280f88","Type":"ContainerDied","Data":"d2a30c3ea32e8635efa0899c2a6cb7d363948b4efbea4945f4879beb8316ce3f"} Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.932349 4945 scope.go:117] "RemoveContainer" containerID="59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca" Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.932461 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.951152 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae81785-b556-4bd4-8a97-c0c5d9a211a6","Type":"ContainerStarted","Data":"cc33750062c40ad3dcd32375b0d9aa0c6c544e8805e5e243c92807b274f7de98"} Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.981546 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b83cf13e-faa9-46d6-a44e-560c56280f88-logs\") pod \"b83cf13e-faa9-46d6-a44e-560c56280f88\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.981761 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn5cr\" (UniqueName: \"kubernetes.io/projected/b83cf13e-faa9-46d6-a44e-560c56280f88-kube-api-access-xn5cr\") pod \"b83cf13e-faa9-46d6-a44e-560c56280f88\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.981824 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b83cf13e-faa9-46d6-a44e-560c56280f88-combined-ca-bundle\") pod \"b83cf13e-faa9-46d6-a44e-560c56280f88\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.981911 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b83cf13e-faa9-46d6-a44e-560c56280f88-config-data\") pod \"b83cf13e-faa9-46d6-a44e-560c56280f88\" (UID: \"b83cf13e-faa9-46d6-a44e-560c56280f88\") " Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.982208 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b83cf13e-faa9-46d6-a44e-560c56280f88-logs" (OuterVolumeSpecName: "logs") pod "b83cf13e-faa9-46d6-a44e-560c56280f88" (UID: "b83cf13e-faa9-46d6-a44e-560c56280f88"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.982592 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b83cf13e-faa9-46d6-a44e-560c56280f88-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:27 crc kubenswrapper[4945]: I1014 15:35:27.997993 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b83cf13e-faa9-46d6-a44e-560c56280f88-kube-api-access-xn5cr" (OuterVolumeSpecName: "kube-api-access-xn5cr") pod "b83cf13e-faa9-46d6-a44e-560c56280f88" (UID: "b83cf13e-faa9-46d6-a44e-560c56280f88"). InnerVolumeSpecName "kube-api-access-xn5cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.029115 4945 scope.go:117] "RemoveContainer" containerID="a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.035365 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b83cf13e-faa9-46d6-a44e-560c56280f88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b83cf13e-faa9-46d6-a44e-560c56280f88" (UID: "b83cf13e-faa9-46d6-a44e-560c56280f88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.038991 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b83cf13e-faa9-46d6-a44e-560c56280f88-config-data" (OuterVolumeSpecName: "config-data") pod "b83cf13e-faa9-46d6-a44e-560c56280f88" (UID: "b83cf13e-faa9-46d6-a44e-560c56280f88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.084015 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b83cf13e-faa9-46d6-a44e-560c56280f88-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.084059 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn5cr\" (UniqueName: \"kubernetes.io/projected/b83cf13e-faa9-46d6-a44e-560c56280f88-kube-api-access-xn5cr\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.084073 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b83cf13e-faa9-46d6-a44e-560c56280f88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.093367 4945 scope.go:117] "RemoveContainer" containerID="59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca" Oct 14 15:35:28 crc kubenswrapper[4945]: E1014 15:35:28.099577 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca\": container with ID starting with 59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca not found: ID does not exist" containerID="59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.099627 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca"} err="failed to get container status \"59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca\": rpc error: code = NotFound desc = could not find container \"59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca\": container with ID starting with 59028cc2033406c84415d8a9d4aa19a6108adc6a86e738fc3770748765be89ca not found: ID does not exist" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.099654 4945 scope.go:117] "RemoveContainer" containerID="a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97" Oct 14 15:35:28 crc kubenswrapper[4945]: E1014 15:35:28.109755 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97\": container with ID starting with a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97 not found: ID does not exist" containerID="a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.109805 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97"} err="failed to get container status \"a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97\": rpc error: code = NotFound desc = could not find container \"a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97\": container with ID starting with a9e7492ce4bfd25000c4b48e498f8b60e3a21d752f8a2c3684fc99fa2c4ecd97 not found: ID does not exist" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.274236 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.286501 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.307013 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:28 crc kubenswrapper[4945]: E1014 15:35:28.307516 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b83cf13e-faa9-46d6-a44e-560c56280f88" containerName="nova-api-api" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.307539 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b83cf13e-faa9-46d6-a44e-560c56280f88" containerName="nova-api-api" Oct 14 15:35:28 crc kubenswrapper[4945]: E1014 15:35:28.307561 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b83cf13e-faa9-46d6-a44e-560c56280f88" containerName="nova-api-log" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.307569 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b83cf13e-faa9-46d6-a44e-560c56280f88" containerName="nova-api-log" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.307800 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b83cf13e-faa9-46d6-a44e-560c56280f88" containerName="nova-api-log" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.307849 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b83cf13e-faa9-46d6-a44e-560c56280f88" containerName="nova-api-api" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.309950 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.311967 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.312176 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.312345 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.324860 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.389457 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.389759 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-config-data\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.389941 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.390043 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e45a0285-7c45-43c2-a711-c6f792c039a3-logs\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.390236 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-public-tls-certs\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.390381 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlnnd\" (UniqueName: \"kubernetes.io/projected/e45a0285-7c45-43c2-a711-c6f792c039a3-kube-api-access-rlnnd\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.491446 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-public-tls-certs\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.491502 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlnnd\" (UniqueName: \"kubernetes.io/projected/e45a0285-7c45-43c2-a711-c6f792c039a3-kube-api-access-rlnnd\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.491799 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.491967 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-config-data\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.492024 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.492059 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e45a0285-7c45-43c2-a711-c6f792c039a3-logs\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.492560 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e45a0285-7c45-43c2-a711-c6f792c039a3-logs\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.497201 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.497438 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-config-data\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.497201 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-public-tls-certs\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.507389 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.509846 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlnnd\" (UniqueName: \"kubernetes.io/projected/e45a0285-7c45-43c2-a711-c6f792c039a3-kube-api-access-rlnnd\") pod \"nova-api-0\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.635400 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.784304 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b83cf13e-faa9-46d6-a44e-560c56280f88" path="/var/lib/kubelet/pods/b83cf13e-faa9-46d6-a44e-560c56280f88/volumes" Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.971159 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae81785-b556-4bd4-8a97-c0c5d9a211a6","Type":"ContainerStarted","Data":"e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f"} Oct 14 15:35:28 crc kubenswrapper[4945]: I1014 15:35:28.971206 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae81785-b556-4bd4-8a97-c0c5d9a211a6","Type":"ContainerStarted","Data":"35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890"} Oct 14 15:35:29 crc kubenswrapper[4945]: I1014 15:35:29.137829 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:29 crc kubenswrapper[4945]: W1014 15:35:29.142036 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode45a0285_7c45_43c2_a711_c6f792c039a3.slice/crio-a0318101913db3a867dae403467eeea82bd172bd58934efdb4ed60fe7af5222a WatchSource:0}: Error finding container a0318101913db3a867dae403467eeea82bd172bd58934efdb4ed60fe7af5222a: Status 404 returned error can't find the container with id a0318101913db3a867dae403467eeea82bd172bd58934efdb4ed60fe7af5222a Oct 14 15:35:29 crc kubenswrapper[4945]: I1014 15:35:29.983698 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae81785-b556-4bd4-8a97-c0c5d9a211a6","Type":"ContainerStarted","Data":"4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee"} Oct 14 15:35:29 crc kubenswrapper[4945]: I1014 15:35:29.985894 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e45a0285-7c45-43c2-a711-c6f792c039a3","Type":"ContainerStarted","Data":"79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5"} Oct 14 15:35:29 crc kubenswrapper[4945]: I1014 15:35:29.985929 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e45a0285-7c45-43c2-a711-c6f792c039a3","Type":"ContainerStarted","Data":"5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696"} Oct 14 15:35:29 crc kubenswrapper[4945]: I1014 15:35:29.985944 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e45a0285-7c45-43c2-a711-c6f792c039a3","Type":"ContainerStarted","Data":"a0318101913db3a867dae403467eeea82bd172bd58934efdb4ed60fe7af5222a"} Oct 14 15:35:30 crc kubenswrapper[4945]: I1014 15:35:30.009221 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.009198019 podStartE2EDuration="2.009198019s" podCreationTimestamp="2025-10-14 15:35:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:35:30.005218355 +0000 UTC m=+1259.989266723" watchObservedRunningTime="2025-10-14 15:35:30.009198019 +0000 UTC m=+1259.993246387" Oct 14 15:35:31 crc kubenswrapper[4945]: I1014 15:35:31.305260 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:31 crc kubenswrapper[4945]: I1014 15:35:31.376904 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:31 crc kubenswrapper[4945]: I1014 15:35:31.519005 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c7b6c5df9-hhtr4" Oct 14 15:35:31 crc kubenswrapper[4945]: I1014 15:35:31.589490 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-hw5kn"] Oct 14 15:35:31 crc kubenswrapper[4945]: I1014 15:35:31.589758 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" podUID="fbbb037f-601a-4b65-8d5e-cd0a57e9957b" containerName="dnsmasq-dns" containerID="cri-o://ce4a3c3cedfd690774e20edda998c949e732f95fa275b6d433dd83fee61c0f76" gracePeriod=10 Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.013671 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae81785-b556-4bd4-8a97-c0c5d9a211a6","Type":"ContainerStarted","Data":"edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064"} Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.014230 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.013957 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="ceilometer-central-agent" containerID="cri-o://35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890" gracePeriod=30 Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.014480 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="sg-core" containerID="cri-o://4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee" gracePeriod=30 Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.014557 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="ceilometer-notification-agent" containerID="cri-o://e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f" gracePeriod=30 Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.014473 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="proxy-httpd" containerID="cri-o://edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064" gracePeriod=30 Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.024170 4945 generic.go:334] "Generic (PLEG): container finished" podID="fbbb037f-601a-4b65-8d5e-cd0a57e9957b" containerID="ce4a3c3cedfd690774e20edda998c949e732f95fa275b6d433dd83fee61c0f76" exitCode=0 Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.024770 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" event={"ID":"fbbb037f-601a-4b65-8d5e-cd0a57e9957b","Type":"ContainerDied","Data":"ce4a3c3cedfd690774e20edda998c949e732f95fa275b6d433dd83fee61c0f76"} Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.024818 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" event={"ID":"fbbb037f-601a-4b65-8d5e-cd0a57e9957b","Type":"ContainerDied","Data":"02f423e833cad5d6ca98a00221bf94092334ecbb5f6bc685f0cb210f7a4e2fec"} Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.024829 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02f423e833cad5d6ca98a00221bf94092334ecbb5f6bc685f0cb210f7a4e2fec" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.045119 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.068955956 podStartE2EDuration="6.045097631s" podCreationTimestamp="2025-10-14 15:35:26 +0000 UTC" firstStartedPulling="2025-10-14 15:35:27.106014981 +0000 UTC m=+1257.090063369" lastFinishedPulling="2025-10-14 15:35:31.082156676 +0000 UTC m=+1261.066205044" observedRunningTime="2025-10-14 15:35:32.039958655 +0000 UTC m=+1262.024007033" watchObservedRunningTime="2025-10-14 15:35:32.045097631 +0000 UTC m=+1262.029145999" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.056405 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.203636 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.225540 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-nvcm2"] Oct 14 15:35:32 crc kubenswrapper[4945]: E1014 15:35:32.225945 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbbb037f-601a-4b65-8d5e-cd0a57e9957b" containerName="init" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.225965 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbbb037f-601a-4b65-8d5e-cd0a57e9957b" containerName="init" Oct 14 15:35:32 crc kubenswrapper[4945]: E1014 15:35:32.225974 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbbb037f-601a-4b65-8d5e-cd0a57e9957b" containerName="dnsmasq-dns" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.225981 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbbb037f-601a-4b65-8d5e-cd0a57e9957b" containerName="dnsmasq-dns" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.226172 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbbb037f-601a-4b65-8d5e-cd0a57e9957b" containerName="dnsmasq-dns" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.226825 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.232998 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.233203 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.242311 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-nvcm2"] Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.290530 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gc9b\" (UniqueName: \"kubernetes.io/projected/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-kube-api-access-7gc9b\") pod \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.290620 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-dns-svc\") pod \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.290776 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-config\") pod \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.290823 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-dns-swift-storage-0\") pod \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.291097 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-ovsdbserver-nb\") pod \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.291171 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-ovsdbserver-sb\") pod \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.291565 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-nvcm2\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.291659 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-scripts\") pod \"nova-cell1-cell-mapping-nvcm2\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.291728 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-config-data\") pod \"nova-cell1-cell-mapping-nvcm2\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.291791 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz965\" (UniqueName: \"kubernetes.io/projected/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-kube-api-access-zz965\") pod \"nova-cell1-cell-mapping-nvcm2\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.316072 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-kube-api-access-7gc9b" (OuterVolumeSpecName: "kube-api-access-7gc9b") pod "fbbb037f-601a-4b65-8d5e-cd0a57e9957b" (UID: "fbbb037f-601a-4b65-8d5e-cd0a57e9957b"). InnerVolumeSpecName "kube-api-access-7gc9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.392588 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-config-data\") pod \"nova-cell1-cell-mapping-nvcm2\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.392650 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz965\" (UniqueName: \"kubernetes.io/projected/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-kube-api-access-zz965\") pod \"nova-cell1-cell-mapping-nvcm2\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.392721 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-nvcm2\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.392765 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-scripts\") pod \"nova-cell1-cell-mapping-nvcm2\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.392809 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gc9b\" (UniqueName: \"kubernetes.io/projected/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-kube-api-access-7gc9b\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.394596 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fbbb037f-601a-4b65-8d5e-cd0a57e9957b" (UID: "fbbb037f-601a-4b65-8d5e-cd0a57e9957b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.395823 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fbbb037f-601a-4b65-8d5e-cd0a57e9957b" (UID: "fbbb037f-601a-4b65-8d5e-cd0a57e9957b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.403713 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-scripts\") pod \"nova-cell1-cell-mapping-nvcm2\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.414555 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-nvcm2\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.415227 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-config-data\") pod \"nova-cell1-cell-mapping-nvcm2\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.458545 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz965\" (UniqueName: \"kubernetes.io/projected/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-kube-api-access-zz965\") pod \"nova-cell1-cell-mapping-nvcm2\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.484919 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fbbb037f-601a-4b65-8d5e-cd0a57e9957b" (UID: "fbbb037f-601a-4b65-8d5e-cd0a57e9957b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.485429 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fbbb037f-601a-4b65-8d5e-cd0a57e9957b" (UID: "fbbb037f-601a-4b65-8d5e-cd0a57e9957b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.493443 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-config" (OuterVolumeSpecName: "config") pod "fbbb037f-601a-4b65-8d5e-cd0a57e9957b" (UID: "fbbb037f-601a-4b65-8d5e-cd0a57e9957b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.493572 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-config\") pod \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\" (UID: \"fbbb037f-601a-4b65-8d5e-cd0a57e9957b\") " Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.494130 4945 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.494149 4945 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.494158 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.494171 4945 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:32 crc kubenswrapper[4945]: W1014 15:35:32.494181 4945 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/fbbb037f-601a-4b65-8d5e-cd0a57e9957b/volumes/kubernetes.io~configmap/config Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.494220 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-config" (OuterVolumeSpecName: "config") pod "fbbb037f-601a-4b65-8d5e-cd0a57e9957b" (UID: "fbbb037f-601a-4b65-8d5e-cd0a57e9957b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.578406 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:32 crc kubenswrapper[4945]: I1014 15:35:32.596300 4945 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbbb037f-601a-4b65-8d5e-cd0a57e9957b-config\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.029709 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-nvcm2"] Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.039470 4945 generic.go:334] "Generic (PLEG): container finished" podID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerID="edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064" exitCode=0 Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.039509 4945 generic.go:334] "Generic (PLEG): container finished" podID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerID="4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee" exitCode=2 Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.039520 4945 generic.go:334] "Generic (PLEG): container finished" podID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerID="e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f" exitCode=0 Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.039584 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.040307 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae81785-b556-4bd4-8a97-c0c5d9a211a6","Type":"ContainerDied","Data":"edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064"} Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.040335 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae81785-b556-4bd4-8a97-c0c5d9a211a6","Type":"ContainerDied","Data":"4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee"} Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.040346 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae81785-b556-4bd4-8a97-c0c5d9a211a6","Type":"ContainerDied","Data":"e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f"} Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.155845 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-hw5kn"] Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.159927 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-hw5kn"] Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.426515 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.529952 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae81785-b556-4bd4-8a97-c0c5d9a211a6-run-httpd\") pod \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.530046 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-scripts\") pod \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.530099 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-ceilometer-tls-certs\") pod \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.530146 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-config-data\") pod \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.530272 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae81785-b556-4bd4-8a97-c0c5d9a211a6-log-httpd\") pod \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.530309 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-sg-core-conf-yaml\") pod \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.530364 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aae81785-b556-4bd4-8a97-c0c5d9a211a6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aae81785-b556-4bd4-8a97-c0c5d9a211a6" (UID: "aae81785-b556-4bd4-8a97-c0c5d9a211a6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.530380 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-combined-ca-bundle\") pod \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.530416 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5fsx\" (UniqueName: \"kubernetes.io/projected/aae81785-b556-4bd4-8a97-c0c5d9a211a6-kube-api-access-b5fsx\") pod \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\" (UID: \"aae81785-b556-4bd4-8a97-c0c5d9a211a6\") " Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.531068 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aae81785-b556-4bd4-8a97-c0c5d9a211a6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aae81785-b556-4bd4-8a97-c0c5d9a211a6" (UID: "aae81785-b556-4bd4-8a97-c0c5d9a211a6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.531384 4945 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae81785-b556-4bd4-8a97-c0c5d9a211a6-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.531408 4945 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae81785-b556-4bd4-8a97-c0c5d9a211a6-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.534271 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae81785-b556-4bd4-8a97-c0c5d9a211a6-kube-api-access-b5fsx" (OuterVolumeSpecName: "kube-api-access-b5fsx") pod "aae81785-b556-4bd4-8a97-c0c5d9a211a6" (UID: "aae81785-b556-4bd4-8a97-c0c5d9a211a6"). InnerVolumeSpecName "kube-api-access-b5fsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.535704 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-scripts" (OuterVolumeSpecName: "scripts") pod "aae81785-b556-4bd4-8a97-c0c5d9a211a6" (UID: "aae81785-b556-4bd4-8a97-c0c5d9a211a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.563008 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aae81785-b556-4bd4-8a97-c0c5d9a211a6" (UID: "aae81785-b556-4bd4-8a97-c0c5d9a211a6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.587045 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "aae81785-b556-4bd4-8a97-c0c5d9a211a6" (UID: "aae81785-b556-4bd4-8a97-c0c5d9a211a6"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.615442 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aae81785-b556-4bd4-8a97-c0c5d9a211a6" (UID: "aae81785-b556-4bd4-8a97-c0c5d9a211a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.634015 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.634091 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5fsx\" (UniqueName: \"kubernetes.io/projected/aae81785-b556-4bd4-8a97-c0c5d9a211a6-kube-api-access-b5fsx\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.634191 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.634209 4945 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.634222 4945 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.637029 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-config-data" (OuterVolumeSpecName: "config-data") pod "aae81785-b556-4bd4-8a97-c0c5d9a211a6" (UID: "aae81785-b556-4bd4-8a97-c0c5d9a211a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:33 crc kubenswrapper[4945]: I1014 15:35:33.736071 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae81785-b556-4bd4-8a97-c0c5d9a211a6-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.053479 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-nvcm2" event={"ID":"c2518d4a-a272-4e4e-8c9f-8884d8e4f612","Type":"ContainerStarted","Data":"b3109b3cfe6b3623e04b24200bae138dd64b7a099ddd410a12734b43af5f84f2"} Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.053523 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-nvcm2" event={"ID":"c2518d4a-a272-4e4e-8c9f-8884d8e4f612","Type":"ContainerStarted","Data":"98e55bcaf2d53bc540c45d9dc62735f726b3c30cb9166ca3ecb7d1c6720a3927"} Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.057850 4945 generic.go:334] "Generic (PLEG): container finished" podID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerID="35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890" exitCode=0 Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.057932 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae81785-b556-4bd4-8a97-c0c5d9a211a6","Type":"ContainerDied","Data":"35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890"} Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.057983 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae81785-b556-4bd4-8a97-c0c5d9a211a6","Type":"ContainerDied","Data":"cc33750062c40ad3dcd32375b0d9aa0c6c544e8805e5e243c92807b274f7de98"} Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.057924 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.058018 4945 scope.go:117] "RemoveContainer" containerID="edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.081011 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-nvcm2" podStartSLOduration=2.080982895 podStartE2EDuration="2.080982895s" podCreationTimestamp="2025-10-14 15:35:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:35:34.079491852 +0000 UTC m=+1264.063540280" watchObservedRunningTime="2025-10-14 15:35:34.080982895 +0000 UTC m=+1264.065031303" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.085135 4945 scope.go:117] "RemoveContainer" containerID="4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.121522 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.131196 4945 scope.go:117] "RemoveContainer" containerID="e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.166275 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.173831 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:34 crc kubenswrapper[4945]: E1014 15:35:34.174822 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="ceilometer-central-agent" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.174842 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="ceilometer-central-agent" Oct 14 15:35:34 crc kubenswrapper[4945]: E1014 15:35:34.174860 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="sg-core" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.174866 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="sg-core" Oct 14 15:35:34 crc kubenswrapper[4945]: E1014 15:35:34.176953 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="proxy-httpd" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.176981 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="proxy-httpd" Oct 14 15:35:34 crc kubenswrapper[4945]: E1014 15:35:34.177005 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="ceilometer-notification-agent" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.177015 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="ceilometer-notification-agent" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.176642 4945 scope.go:117] "RemoveContainer" containerID="35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.177897 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="sg-core" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.177932 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="ceilometer-central-agent" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.177951 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="proxy-httpd" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.177973 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" containerName="ceilometer-notification-agent" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.182621 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.185435 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.185678 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.190202 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.190320 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.220839 4945 scope.go:117] "RemoveContainer" containerID="edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064" Oct 14 15:35:34 crc kubenswrapper[4945]: E1014 15:35:34.221509 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064\": container with ID starting with edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064 not found: ID does not exist" containerID="edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.221607 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064"} err="failed to get container status \"edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064\": rpc error: code = NotFound desc = could not find container \"edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064\": container with ID starting with edf493048a74640e33ec0cac59157de9f882f31680dd684a456905d247003064 not found: ID does not exist" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.221711 4945 scope.go:117] "RemoveContainer" containerID="4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee" Oct 14 15:35:34 crc kubenswrapper[4945]: E1014 15:35:34.222213 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee\": container with ID starting with 4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee not found: ID does not exist" containerID="4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.222245 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee"} err="failed to get container status \"4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee\": rpc error: code = NotFound desc = could not find container \"4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee\": container with ID starting with 4d96f8402f87b911a54f0c80f90884f9a7adc4d43aea0114e5587de7dc126aee not found: ID does not exist" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.222269 4945 scope.go:117] "RemoveContainer" containerID="e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f" Oct 14 15:35:34 crc kubenswrapper[4945]: E1014 15:35:34.222488 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f\": container with ID starting with e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f not found: ID does not exist" containerID="e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.222508 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f"} err="failed to get container status \"e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f\": rpc error: code = NotFound desc = could not find container \"e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f\": container with ID starting with e394e0e27196a5592a6d4acb8537299008f622279a5b44c9c310acf69e6e5e5f not found: ID does not exist" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.222522 4945 scope.go:117] "RemoveContainer" containerID="35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890" Oct 14 15:35:34 crc kubenswrapper[4945]: E1014 15:35:34.222779 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890\": container with ID starting with 35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890 not found: ID does not exist" containerID="35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.222861 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890"} err="failed to get container status \"35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890\": rpc error: code = NotFound desc = could not find container \"35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890\": container with ID starting with 35121eb9de53011b25f801bc6e6938a7d56deacfe66bc9dbd1fc381fdc1e3890 not found: ID does not exist" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.274113 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.274159 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/519a62f5-6d77-4b41-9d34-52651d4e9571-log-httpd\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.274183 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfmw6\" (UniqueName: \"kubernetes.io/projected/519a62f5-6d77-4b41-9d34-52651d4e9571-kube-api-access-lfmw6\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.274199 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-scripts\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.274238 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/519a62f5-6d77-4b41-9d34-52651d4e9571-run-httpd\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.274263 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-config-data\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.274294 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.274325 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.375685 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.375752 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.375851 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.375870 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/519a62f5-6d77-4b41-9d34-52651d4e9571-log-httpd\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.375916 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfmw6\" (UniqueName: \"kubernetes.io/projected/519a62f5-6d77-4b41-9d34-52651d4e9571-kube-api-access-lfmw6\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.375939 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-scripts\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.375989 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/519a62f5-6d77-4b41-9d34-52651d4e9571-run-httpd\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.376022 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-config-data\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.378001 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/519a62f5-6d77-4b41-9d34-52651d4e9571-run-httpd\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.378010 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/519a62f5-6d77-4b41-9d34-52651d4e9571-log-httpd\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.380763 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-config-data\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.382782 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.385782 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.398451 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.400684 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/519a62f5-6d77-4b41-9d34-52651d4e9571-scripts\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.401568 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfmw6\" (UniqueName: \"kubernetes.io/projected/519a62f5-6d77-4b41-9d34-52651d4e9571-kube-api-access-lfmw6\") pod \"ceilometer-0\" (UID: \"519a62f5-6d77-4b41-9d34-52651d4e9571\") " pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.500543 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.781299 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aae81785-b556-4bd4-8a97-c0c5d9a211a6" path="/var/lib/kubelet/pods/aae81785-b556-4bd4-8a97-c0c5d9a211a6/volumes" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.783831 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbbb037f-601a-4b65-8d5e-cd0a57e9957b" path="/var/lib/kubelet/pods/fbbb037f-601a-4b65-8d5e-cd0a57e9957b/volumes" Oct 14 15:35:34 crc kubenswrapper[4945]: I1014 15:35:34.956504 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 15:35:34 crc kubenswrapper[4945]: W1014 15:35:34.962706 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod519a62f5_6d77_4b41_9d34_52651d4e9571.slice/crio-9ab66fbd01be9f5492f67ca9ddea4c3b49d34a0782784cb9a0f4973cccc11eeb WatchSource:0}: Error finding container 9ab66fbd01be9f5492f67ca9ddea4c3b49d34a0782784cb9a0f4973cccc11eeb: Status 404 returned error can't find the container with id 9ab66fbd01be9f5492f67ca9ddea4c3b49d34a0782784cb9a0f4973cccc11eeb Oct 14 15:35:35 crc kubenswrapper[4945]: I1014 15:35:35.080116 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"519a62f5-6d77-4b41-9d34-52651d4e9571","Type":"ContainerStarted","Data":"9ab66fbd01be9f5492f67ca9ddea4c3b49d34a0782784cb9a0f4973cccc11eeb"} Oct 14 15:35:36 crc kubenswrapper[4945]: I1014 15:35:36.098480 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"519a62f5-6d77-4b41-9d34-52651d4e9571","Type":"ContainerStarted","Data":"48de943df7e23f17699204ee8dd416ad4a1f748cd0f0ab5469ba11f732188a77"} Oct 14 15:35:36 crc kubenswrapper[4945]: I1014 15:35:36.819781 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-865f5d856f-hw5kn" podUID="fbbb037f-601a-4b65-8d5e-cd0a57e9957b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.196:5353: i/o timeout" Oct 14 15:35:37 crc kubenswrapper[4945]: I1014 15:35:37.112666 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"519a62f5-6d77-4b41-9d34-52651d4e9571","Type":"ContainerStarted","Data":"71c80dd67cc2107022b83cd8bb13f8ef797d55d0cecba2ced377a058d12e5687"} Oct 14 15:35:38 crc kubenswrapper[4945]: I1014 15:35:38.124797 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"519a62f5-6d77-4b41-9d34-52651d4e9571","Type":"ContainerStarted","Data":"9f0c4292b3bfed7098f1796dd97f7c47563b9477333a6e38760d17875f582644"} Oct 14 15:35:38 crc kubenswrapper[4945]: I1014 15:35:38.635828 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 15:35:38 crc kubenswrapper[4945]: I1014 15:35:38.636655 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 15:35:39 crc kubenswrapper[4945]: I1014 15:35:39.135727 4945 generic.go:334] "Generic (PLEG): container finished" podID="c2518d4a-a272-4e4e-8c9f-8884d8e4f612" containerID="b3109b3cfe6b3623e04b24200bae138dd64b7a099ddd410a12734b43af5f84f2" exitCode=0 Oct 14 15:35:39 crc kubenswrapper[4945]: I1014 15:35:39.135783 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-nvcm2" event={"ID":"c2518d4a-a272-4e4e-8c9f-8884d8e4f612","Type":"ContainerDied","Data":"b3109b3cfe6b3623e04b24200bae138dd64b7a099ddd410a12734b43af5f84f2"} Oct 14 15:35:39 crc kubenswrapper[4945]: I1014 15:35:39.652150 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e45a0285-7c45-43c2-a711-c6f792c039a3" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 15:35:39 crc kubenswrapper[4945]: I1014 15:35:39.652141 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e45a0285-7c45-43c2-a711-c6f792c039a3" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.148890 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"519a62f5-6d77-4b41-9d34-52651d4e9571","Type":"ContainerStarted","Data":"0bceb8f9fe0683ce08bd727d99454147b1d5d6b298d9643c6a1436313fe87200"} Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.149126 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.175710 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9568908889999999 podStartE2EDuration="6.175680812s" podCreationTimestamp="2025-10-14 15:35:34 +0000 UTC" firstStartedPulling="2025-10-14 15:35:34.966228082 +0000 UTC m=+1264.950276450" lastFinishedPulling="2025-10-14 15:35:39.185018005 +0000 UTC m=+1269.169066373" observedRunningTime="2025-10-14 15:35:40.169491236 +0000 UTC m=+1270.153539594" watchObservedRunningTime="2025-10-14 15:35:40.175680812 +0000 UTC m=+1270.159729180" Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.599448 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.720991 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-config-data\") pod \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.721324 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz965\" (UniqueName: \"kubernetes.io/projected/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-kube-api-access-zz965\") pod \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.721472 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-combined-ca-bundle\") pod \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.721631 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-scripts\") pod \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\" (UID: \"c2518d4a-a272-4e4e-8c9f-8884d8e4f612\") " Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.730057 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-scripts" (OuterVolumeSpecName: "scripts") pod "c2518d4a-a272-4e4e-8c9f-8884d8e4f612" (UID: "c2518d4a-a272-4e4e-8c9f-8884d8e4f612"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.734084 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-kube-api-access-zz965" (OuterVolumeSpecName: "kube-api-access-zz965") pod "c2518d4a-a272-4e4e-8c9f-8884d8e4f612" (UID: "c2518d4a-a272-4e4e-8c9f-8884d8e4f612"). InnerVolumeSpecName "kube-api-access-zz965". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.757034 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2518d4a-a272-4e4e-8c9f-8884d8e4f612" (UID: "c2518d4a-a272-4e4e-8c9f-8884d8e4f612"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.757174 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-config-data" (OuterVolumeSpecName: "config-data") pod "c2518d4a-a272-4e4e-8c9f-8884d8e4f612" (UID: "c2518d4a-a272-4e4e-8c9f-8884d8e4f612"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.825373 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.825708 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz965\" (UniqueName: \"kubernetes.io/projected/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-kube-api-access-zz965\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.825719 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:40 crc kubenswrapper[4945]: I1014 15:35:40.825730 4945 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2518d4a-a272-4e4e-8c9f-8884d8e4f612-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:41 crc kubenswrapper[4945]: I1014 15:35:41.160243 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-nvcm2" event={"ID":"c2518d4a-a272-4e4e-8c9f-8884d8e4f612","Type":"ContainerDied","Data":"98e55bcaf2d53bc540c45d9dc62735f726b3c30cb9166ca3ecb7d1c6720a3927"} Oct 14 15:35:41 crc kubenswrapper[4945]: I1014 15:35:41.160278 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-nvcm2" Oct 14 15:35:41 crc kubenswrapper[4945]: I1014 15:35:41.160289 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98e55bcaf2d53bc540c45d9dc62735f726b3c30cb9166ca3ecb7d1c6720a3927" Oct 14 15:35:41 crc kubenswrapper[4945]: I1014 15:35:41.356684 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:41 crc kubenswrapper[4945]: I1014 15:35:41.356901 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e45a0285-7c45-43c2-a711-c6f792c039a3" containerName="nova-api-log" containerID="cri-o://5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696" gracePeriod=30 Oct 14 15:35:41 crc kubenswrapper[4945]: I1014 15:35:41.356960 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e45a0285-7c45-43c2-a711-c6f792c039a3" containerName="nova-api-api" containerID="cri-o://79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5" gracePeriod=30 Oct 14 15:35:41 crc kubenswrapper[4945]: I1014 15:35:41.389805 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:35:41 crc kubenswrapper[4945]: I1014 15:35:41.390605 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d65e47e7-2cb8-492c-aef0-6bc069d1f643" containerName="nova-scheduler-scheduler" containerID="cri-o://6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4" gracePeriod=30 Oct 14 15:35:41 crc kubenswrapper[4945]: I1014 15:35:41.414523 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:35:41 crc kubenswrapper[4945]: I1014 15:35:41.414806 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5e25b244-f49c-43e0-9cdd-395f0906c15c" containerName="nova-metadata-log" containerID="cri-o://6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64" gracePeriod=30 Oct 14 15:35:41 crc kubenswrapper[4945]: I1014 15:35:41.415012 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5e25b244-f49c-43e0-9cdd-395f0906c15c" containerName="nova-metadata-metadata" containerID="cri-o://837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622" gracePeriod=30 Oct 14 15:35:41 crc kubenswrapper[4945]: E1014 15:35:41.588546 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode45a0285_7c45_43c2_a711_c6f792c039a3.slice/crio-conmon-5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e25b244_f49c_43e0_9cdd_395f0906c15c.slice/crio-6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e25b244_f49c_43e0_9cdd_395f0906c15c.slice/crio-conmon-6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode45a0285_7c45_43c2_a711_c6f792c039a3.slice/crio-5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696.scope\": RecentStats: unable to find data in memory cache]" Oct 14 15:35:42 crc kubenswrapper[4945]: I1014 15:35:42.171611 4945 generic.go:334] "Generic (PLEG): container finished" podID="5e25b244-f49c-43e0-9cdd-395f0906c15c" containerID="6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64" exitCode=143 Oct 14 15:35:42 crc kubenswrapper[4945]: I1014 15:35:42.172042 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e25b244-f49c-43e0-9cdd-395f0906c15c","Type":"ContainerDied","Data":"6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64"} Oct 14 15:35:42 crc kubenswrapper[4945]: I1014 15:35:42.175199 4945 generic.go:334] "Generic (PLEG): container finished" podID="e45a0285-7c45-43c2-a711-c6f792c039a3" containerID="5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696" exitCode=143 Oct 14 15:35:42 crc kubenswrapper[4945]: I1014 15:35:42.175269 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e45a0285-7c45-43c2-a711-c6f792c039a3","Type":"ContainerDied","Data":"5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696"} Oct 14 15:35:42 crc kubenswrapper[4945]: E1014 15:35:42.924228 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 15:35:42 crc kubenswrapper[4945]: E1014 15:35:42.926216 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 15:35:42 crc kubenswrapper[4945]: E1014 15:35:42.927760 4945 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 15:35:42 crc kubenswrapper[4945]: E1014 15:35:42.927835 4945 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="d65e47e7-2cb8-492c-aef0-6bc069d1f643" containerName="nova-scheduler-scheduler" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.077564 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.084353 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.208424 4945 generic.go:334] "Generic (PLEG): container finished" podID="e45a0285-7c45-43c2-a711-c6f792c039a3" containerID="79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5" exitCode=0 Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.208490 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e45a0285-7c45-43c2-a711-c6f792c039a3","Type":"ContainerDied","Data":"79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5"} Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.208516 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e45a0285-7c45-43c2-a711-c6f792c039a3","Type":"ContainerDied","Data":"a0318101913db3a867dae403467eeea82bd172bd58934efdb4ed60fe7af5222a"} Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.208533 4945 scope.go:117] "RemoveContainer" containerID="79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.208675 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.212679 4945 generic.go:334] "Generic (PLEG): container finished" podID="5e25b244-f49c-43e0-9cdd-395f0906c15c" containerID="837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622" exitCode=0 Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.213108 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e25b244-f49c-43e0-9cdd-395f0906c15c","Type":"ContainerDied","Data":"837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622"} Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.213152 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e25b244-f49c-43e0-9cdd-395f0906c15c","Type":"ContainerDied","Data":"2d67eb56b6cfe962edfb78bb95b9ed5aec0e93830364aed13e86d0dbddf6fb19"} Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.213218 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.215245 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-config-data\") pod \"5e25b244-f49c-43e0-9cdd-395f0906c15c\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.215325 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-combined-ca-bundle\") pod \"5e25b244-f49c-43e0-9cdd-395f0906c15c\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.215360 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-combined-ca-bundle\") pod \"e45a0285-7c45-43c2-a711-c6f792c039a3\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.215394 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-nova-metadata-tls-certs\") pod \"5e25b244-f49c-43e0-9cdd-395f0906c15c\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.215479 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-config-data\") pod \"e45a0285-7c45-43c2-a711-c6f792c039a3\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.215510 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24xhb\" (UniqueName: \"kubernetes.io/projected/5e25b244-f49c-43e0-9cdd-395f0906c15c-kube-api-access-24xhb\") pod \"5e25b244-f49c-43e0-9cdd-395f0906c15c\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.215561 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlnnd\" (UniqueName: \"kubernetes.io/projected/e45a0285-7c45-43c2-a711-c6f792c039a3-kube-api-access-rlnnd\") pod \"e45a0285-7c45-43c2-a711-c6f792c039a3\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.215670 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e45a0285-7c45-43c2-a711-c6f792c039a3-logs\") pod \"e45a0285-7c45-43c2-a711-c6f792c039a3\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.215695 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e25b244-f49c-43e0-9cdd-395f0906c15c-logs\") pod \"5e25b244-f49c-43e0-9cdd-395f0906c15c\" (UID: \"5e25b244-f49c-43e0-9cdd-395f0906c15c\") " Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.215719 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-public-tls-certs\") pod \"e45a0285-7c45-43c2-a711-c6f792c039a3\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.215752 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-internal-tls-certs\") pod \"e45a0285-7c45-43c2-a711-c6f792c039a3\" (UID: \"e45a0285-7c45-43c2-a711-c6f792c039a3\") " Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.221970 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e45a0285-7c45-43c2-a711-c6f792c039a3-logs" (OuterVolumeSpecName: "logs") pod "e45a0285-7c45-43c2-a711-c6f792c039a3" (UID: "e45a0285-7c45-43c2-a711-c6f792c039a3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.225867 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e25b244-f49c-43e0-9cdd-395f0906c15c-kube-api-access-24xhb" (OuterVolumeSpecName: "kube-api-access-24xhb") pod "5e25b244-f49c-43e0-9cdd-395f0906c15c" (UID: "5e25b244-f49c-43e0-9cdd-395f0906c15c"). InnerVolumeSpecName "kube-api-access-24xhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.226066 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e25b244-f49c-43e0-9cdd-395f0906c15c-logs" (OuterVolumeSpecName: "logs") pod "5e25b244-f49c-43e0-9cdd-395f0906c15c" (UID: "5e25b244-f49c-43e0-9cdd-395f0906c15c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.251514 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e45a0285-7c45-43c2-a711-c6f792c039a3-kube-api-access-rlnnd" (OuterVolumeSpecName: "kube-api-access-rlnnd") pod "e45a0285-7c45-43c2-a711-c6f792c039a3" (UID: "e45a0285-7c45-43c2-a711-c6f792c039a3"). InnerVolumeSpecName "kube-api-access-rlnnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.254489 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-config-data" (OuterVolumeSpecName: "config-data") pod "5e25b244-f49c-43e0-9cdd-395f0906c15c" (UID: "5e25b244-f49c-43e0-9cdd-395f0906c15c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.254505 4945 scope.go:117] "RemoveContainer" containerID="5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.260136 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e25b244-f49c-43e0-9cdd-395f0906c15c" (UID: "5e25b244-f49c-43e0-9cdd-395f0906c15c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.265841 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e45a0285-7c45-43c2-a711-c6f792c039a3" (UID: "e45a0285-7c45-43c2-a711-c6f792c039a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.269488 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-config-data" (OuterVolumeSpecName: "config-data") pod "e45a0285-7c45-43c2-a711-c6f792c039a3" (UID: "e45a0285-7c45-43c2-a711-c6f792c039a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.288589 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e45a0285-7c45-43c2-a711-c6f792c039a3" (UID: "e45a0285-7c45-43c2-a711-c6f792c039a3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.300083 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e45a0285-7c45-43c2-a711-c6f792c039a3" (UID: "e45a0285-7c45-43c2-a711-c6f792c039a3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.313827 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "5e25b244-f49c-43e0-9cdd-395f0906c15c" (UID: "5e25b244-f49c-43e0-9cdd-395f0906c15c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.318826 4945 scope.go:117] "RemoveContainer" containerID="79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5" Oct 14 15:35:45 crc kubenswrapper[4945]: E1014 15:35:45.319233 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5\": container with ID starting with 79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5 not found: ID does not exist" containerID="79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.319263 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5"} err="failed to get container status \"79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5\": rpc error: code = NotFound desc = could not find container \"79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5\": container with ID starting with 79885a8364a9243e2b109fe48a5126c213c09720acb461ce91c5fe8ab07eb4c5 not found: ID does not exist" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.319283 4945 scope.go:117] "RemoveContainer" containerID="5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696" Oct 14 15:35:45 crc kubenswrapper[4945]: E1014 15:35:45.319559 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696\": container with ID starting with 5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696 not found: ID does not exist" containerID="5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.319594 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696"} err="failed to get container status \"5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696\": rpc error: code = NotFound desc = could not find container \"5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696\": container with ID starting with 5cfc44d02522c6786ae08af5d75741ad18cb7af50b79082e5b0fb6680b554696 not found: ID does not exist" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.319614 4945 scope.go:117] "RemoveContainer" containerID="837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.320299 4945 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.320332 4945 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.320357 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.320370 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.320382 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.320393 4945 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e25b244-f49c-43e0-9cdd-395f0906c15c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.320407 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e45a0285-7c45-43c2-a711-c6f792c039a3-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.320416 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24xhb\" (UniqueName: \"kubernetes.io/projected/5e25b244-f49c-43e0-9cdd-395f0906c15c-kube-api-access-24xhb\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.320424 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlnnd\" (UniqueName: \"kubernetes.io/projected/e45a0285-7c45-43c2-a711-c6f792c039a3-kube-api-access-rlnnd\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.320432 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e45a0285-7c45-43c2-a711-c6f792c039a3-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.320443 4945 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e25b244-f49c-43e0-9cdd-395f0906c15c-logs\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.374181 4945 scope.go:117] "RemoveContainer" containerID="6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.391093 4945 scope.go:117] "RemoveContainer" containerID="837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622" Oct 14 15:35:45 crc kubenswrapper[4945]: E1014 15:35:45.391535 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622\": container with ID starting with 837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622 not found: ID does not exist" containerID="837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.391587 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622"} err="failed to get container status \"837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622\": rpc error: code = NotFound desc = could not find container \"837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622\": container with ID starting with 837524ce3e84a5f57479010149b3518d73e90475e45857ffa3a4be462323f622 not found: ID does not exist" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.391609 4945 scope.go:117] "RemoveContainer" containerID="6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64" Oct 14 15:35:45 crc kubenswrapper[4945]: E1014 15:35:45.392118 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64\": container with ID starting with 6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64 not found: ID does not exist" containerID="6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.392195 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64"} err="failed to get container status \"6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64\": rpc error: code = NotFound desc = could not find container \"6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64\": container with ID starting with 6dcd5a798cd0df4641cca572fe3250c4a4bdf2f6214ad808a294dab656072b64 not found: ID does not exist" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.554051 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.565108 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.575702 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.587946 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.605056 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:45 crc kubenswrapper[4945]: E1014 15:35:45.605461 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e25b244-f49c-43e0-9cdd-395f0906c15c" containerName="nova-metadata-metadata" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.605474 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e25b244-f49c-43e0-9cdd-395f0906c15c" containerName="nova-metadata-metadata" Oct 14 15:35:45 crc kubenswrapper[4945]: E1014 15:35:45.605505 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e45a0285-7c45-43c2-a711-c6f792c039a3" containerName="nova-api-api" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.605512 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e45a0285-7c45-43c2-a711-c6f792c039a3" containerName="nova-api-api" Oct 14 15:35:45 crc kubenswrapper[4945]: E1014 15:35:45.605527 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e45a0285-7c45-43c2-a711-c6f792c039a3" containerName="nova-api-log" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.605533 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="e45a0285-7c45-43c2-a711-c6f792c039a3" containerName="nova-api-log" Oct 14 15:35:45 crc kubenswrapper[4945]: E1014 15:35:45.605543 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e25b244-f49c-43e0-9cdd-395f0906c15c" containerName="nova-metadata-log" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.605549 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e25b244-f49c-43e0-9cdd-395f0906c15c" containerName="nova-metadata-log" Oct 14 15:35:45 crc kubenswrapper[4945]: E1014 15:35:45.605558 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2518d4a-a272-4e4e-8c9f-8884d8e4f612" containerName="nova-manage" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.605564 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2518d4a-a272-4e4e-8c9f-8884d8e4f612" containerName="nova-manage" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.605782 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2518d4a-a272-4e4e-8c9f-8884d8e4f612" containerName="nova-manage" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.605801 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e25b244-f49c-43e0-9cdd-395f0906c15c" containerName="nova-metadata-log" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.605812 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e45a0285-7c45-43c2-a711-c6f792c039a3" containerName="nova-api-log" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.605820 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="e45a0285-7c45-43c2-a711-c6f792c039a3" containerName="nova-api-api" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.605834 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e25b244-f49c-43e0-9cdd-395f0906c15c" containerName="nova-metadata-metadata" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.606812 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.612467 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.612612 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.612694 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.636600 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.646972 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.648938 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.651369 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.655095 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.655744 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.727349 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f22b1bbe-8060-4302-b929-567dc4f33173-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.727414 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkfr9\" (UniqueName: \"kubernetes.io/projected/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-kube-api-access-rkfr9\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.727436 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c42c7\" (UniqueName: \"kubernetes.io/projected/f22b1bbe-8060-4302-b929-567dc4f33173-kube-api-access-c42c7\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.727489 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-config-data\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.727652 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-public-tls-certs\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.727691 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f22b1bbe-8060-4302-b929-567dc4f33173-config-data\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.727721 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f22b1bbe-8060-4302-b929-567dc4f33173-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.727754 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-logs\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.727817 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.727845 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.727889 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f22b1bbe-8060-4302-b929-567dc4f33173-logs\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.829460 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f22b1bbe-8060-4302-b929-567dc4f33173-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.829527 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkfr9\" (UniqueName: \"kubernetes.io/projected/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-kube-api-access-rkfr9\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.829573 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c42c7\" (UniqueName: \"kubernetes.io/projected/f22b1bbe-8060-4302-b929-567dc4f33173-kube-api-access-c42c7\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.829662 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-config-data\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.829750 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-public-tls-certs\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.829777 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f22b1bbe-8060-4302-b929-567dc4f33173-config-data\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.829798 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f22b1bbe-8060-4302-b929-567dc4f33173-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.829825 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-logs\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.829858 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.829899 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.829925 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f22b1bbe-8060-4302-b929-567dc4f33173-logs\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.831842 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-logs\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.832110 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f22b1bbe-8060-4302-b929-567dc4f33173-logs\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.835305 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-public-tls-certs\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.835658 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-config-data\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.835771 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f22b1bbe-8060-4302-b929-567dc4f33173-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.835986 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f22b1bbe-8060-4302-b929-567dc4f33173-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.838350 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.842989 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.851604 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f22b1bbe-8060-4302-b929-567dc4f33173-config-data\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.853324 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkfr9\" (UniqueName: \"kubernetes.io/projected/a821c3a6-1324-4c0e-bcf4-1700cb3cfd69-kube-api-access-rkfr9\") pod \"nova-api-0\" (UID: \"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69\") " pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.857594 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c42c7\" (UniqueName: \"kubernetes.io/projected/f22b1bbe-8060-4302-b929-567dc4f33173-kube-api-access-c42c7\") pod \"nova-metadata-0\" (UID: \"f22b1bbe-8060-4302-b929-567dc4f33173\") " pod="openstack/nova-metadata-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.959953 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 15:35:45 crc kubenswrapper[4945]: I1014 15:35:45.972696 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 15:35:46 crc kubenswrapper[4945]: I1014 15:35:46.785997 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e25b244-f49c-43e0-9cdd-395f0906c15c" path="/var/lib/kubelet/pods/5e25b244-f49c-43e0-9cdd-395f0906c15c/volumes" Oct 14 15:35:46 crc kubenswrapper[4945]: I1014 15:35:46.787288 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e45a0285-7c45-43c2-a711-c6f792c039a3" path="/var/lib/kubelet/pods/e45a0285-7c45-43c2-a711-c6f792c039a3/volumes" Oct 14 15:35:46 crc kubenswrapper[4945]: I1014 15:35:46.795799 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:35:46 crc kubenswrapper[4945]: I1014 15:35:46.795849 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:35:46 crc kubenswrapper[4945]: I1014 15:35:46.795940 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:35:46 crc kubenswrapper[4945]: I1014 15:35:46.797259 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4b33206c225bec396b0367980f7cae9b8b640b794421f9c820e87d051e7c6d3e"} pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 15:35:46 crc kubenswrapper[4945]: I1014 15:35:46.797334 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" containerID="cri-o://4b33206c225bec396b0367980f7cae9b8b640b794421f9c820e87d051e7c6d3e" gracePeriod=600 Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.061644 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.116317 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.129954 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.153572 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65e47e7-2cb8-492c-aef0-6bc069d1f643-config-data\") pod \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\" (UID: \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\") " Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.153898 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8g84\" (UniqueName: \"kubernetes.io/projected/d65e47e7-2cb8-492c-aef0-6bc069d1f643-kube-api-access-t8g84\") pod \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\" (UID: \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\") " Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.154224 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65e47e7-2cb8-492c-aef0-6bc069d1f643-combined-ca-bundle\") pod \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\" (UID: \"d65e47e7-2cb8-492c-aef0-6bc069d1f643\") " Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.170044 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d65e47e7-2cb8-492c-aef0-6bc069d1f643-kube-api-access-t8g84" (OuterVolumeSpecName: "kube-api-access-t8g84") pod "d65e47e7-2cb8-492c-aef0-6bc069d1f643" (UID: "d65e47e7-2cb8-492c-aef0-6bc069d1f643"). InnerVolumeSpecName "kube-api-access-t8g84". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.223423 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65e47e7-2cb8-492c-aef0-6bc069d1f643-config-data" (OuterVolumeSpecName: "config-data") pod "d65e47e7-2cb8-492c-aef0-6bc069d1f643" (UID: "d65e47e7-2cb8-492c-aef0-6bc069d1f643"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.224415 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65e47e7-2cb8-492c-aef0-6bc069d1f643-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d65e47e7-2cb8-492c-aef0-6bc069d1f643" (UID: "d65e47e7-2cb8-492c-aef0-6bc069d1f643"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.238537 4945 generic.go:334] "Generic (PLEG): container finished" podID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerID="4b33206c225bec396b0367980f7cae9b8b640b794421f9c820e87d051e7c6d3e" exitCode=0 Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.238601 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerDied","Data":"4b33206c225bec396b0367980f7cae9b8b640b794421f9c820e87d051e7c6d3e"} Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.238814 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerStarted","Data":"8597fbae8d09b487635545f118989013ee775ca16d722567c5fc29358c69817b"} Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.238909 4945 scope.go:117] "RemoveContainer" containerID="cb33e17badcd091cccc29c2b02c75756661752ad2d76d11d53ff1811c318baac" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.241783 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f22b1bbe-8060-4302-b929-567dc4f33173","Type":"ContainerStarted","Data":"0198fc4c7fe6c58813a7b8db669e6c4625ce67b0f53cc9fcb2ac377f1d280fd2"} Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.243495 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69","Type":"ContainerStarted","Data":"a00d1b2fc2adc9cd3801f147ded7a4a9d4fe2c59756e2b8c0a43ebe244453b5e"} Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.251079 4945 generic.go:334] "Generic (PLEG): container finished" podID="d65e47e7-2cb8-492c-aef0-6bc069d1f643" containerID="6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4" exitCode=0 Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.251119 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d65e47e7-2cb8-492c-aef0-6bc069d1f643","Type":"ContainerDied","Data":"6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4"} Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.251185 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d65e47e7-2cb8-492c-aef0-6bc069d1f643","Type":"ContainerDied","Data":"839af0f67b96485c0f0551546c8e9b2ae6f5f07080d420e0f95e6a2de858dd38"} Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.251246 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.261990 4945 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65e47e7-2cb8-492c-aef0-6bc069d1f643-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.262030 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65e47e7-2cb8-492c-aef0-6bc069d1f643-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.262043 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8g84\" (UniqueName: \"kubernetes.io/projected/d65e47e7-2cb8-492c-aef0-6bc069d1f643-kube-api-access-t8g84\") on node \"crc\" DevicePath \"\"" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.296591 4945 scope.go:117] "RemoveContainer" containerID="6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.304697 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.320622 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.323133 4945 scope.go:117] "RemoveContainer" containerID="6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4" Oct 14 15:35:47 crc kubenswrapper[4945]: E1014 15:35:47.324318 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4\": container with ID starting with 6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4 not found: ID does not exist" containerID="6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.324363 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4"} err="failed to get container status \"6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4\": rpc error: code = NotFound desc = could not find container \"6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4\": container with ID starting with 6090b86abe2e925eb83bad9e7277a4ef5f95cff3a57713c43d85571a81c4cff4 not found: ID does not exist" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.331605 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:35:47 crc kubenswrapper[4945]: E1014 15:35:47.332073 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d65e47e7-2cb8-492c-aef0-6bc069d1f643" containerName="nova-scheduler-scheduler" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.332093 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d65e47e7-2cb8-492c-aef0-6bc069d1f643" containerName="nova-scheduler-scheduler" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.332357 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d65e47e7-2cb8-492c-aef0-6bc069d1f643" containerName="nova-scheduler-scheduler" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.333117 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.336457 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.349892 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.467028 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhb26\" (UniqueName: \"kubernetes.io/projected/524fdd1e-9a42-495f-a692-e4722ab3c4e2-kube-api-access-zhb26\") pod \"nova-scheduler-0\" (UID: \"524fdd1e-9a42-495f-a692-e4722ab3c4e2\") " pod="openstack/nova-scheduler-0" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.467092 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/524fdd1e-9a42-495f-a692-e4722ab3c4e2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"524fdd1e-9a42-495f-a692-e4722ab3c4e2\") " pod="openstack/nova-scheduler-0" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.467198 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/524fdd1e-9a42-495f-a692-e4722ab3c4e2-config-data\") pod \"nova-scheduler-0\" (UID: \"524fdd1e-9a42-495f-a692-e4722ab3c4e2\") " pod="openstack/nova-scheduler-0" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.569420 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhb26\" (UniqueName: \"kubernetes.io/projected/524fdd1e-9a42-495f-a692-e4722ab3c4e2-kube-api-access-zhb26\") pod \"nova-scheduler-0\" (UID: \"524fdd1e-9a42-495f-a692-e4722ab3c4e2\") " pod="openstack/nova-scheduler-0" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.569471 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/524fdd1e-9a42-495f-a692-e4722ab3c4e2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"524fdd1e-9a42-495f-a692-e4722ab3c4e2\") " pod="openstack/nova-scheduler-0" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.569542 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/524fdd1e-9a42-495f-a692-e4722ab3c4e2-config-data\") pod \"nova-scheduler-0\" (UID: \"524fdd1e-9a42-495f-a692-e4722ab3c4e2\") " pod="openstack/nova-scheduler-0" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.574446 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/524fdd1e-9a42-495f-a692-e4722ab3c4e2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"524fdd1e-9a42-495f-a692-e4722ab3c4e2\") " pod="openstack/nova-scheduler-0" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.574609 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/524fdd1e-9a42-495f-a692-e4722ab3c4e2-config-data\") pod \"nova-scheduler-0\" (UID: \"524fdd1e-9a42-495f-a692-e4722ab3c4e2\") " pod="openstack/nova-scheduler-0" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.589365 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhb26\" (UniqueName: \"kubernetes.io/projected/524fdd1e-9a42-495f-a692-e4722ab3c4e2-kube-api-access-zhb26\") pod \"nova-scheduler-0\" (UID: \"524fdd1e-9a42-495f-a692-e4722ab3c4e2\") " pod="openstack/nova-scheduler-0" Oct 14 15:35:47 crc kubenswrapper[4945]: I1014 15:35:47.661242 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 15:35:48 crc kubenswrapper[4945]: W1014 15:35:48.198612 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod524fdd1e_9a42_495f_a692_e4722ab3c4e2.slice/crio-3badb17ed7ebacdbe466dbbd3f9c2765a6194a879f387d6de5553071794a3830 WatchSource:0}: Error finding container 3badb17ed7ebacdbe466dbbd3f9c2765a6194a879f387d6de5553071794a3830: Status 404 returned error can't find the container with id 3badb17ed7ebacdbe466dbbd3f9c2765a6194a879f387d6de5553071794a3830 Oct 14 15:35:48 crc kubenswrapper[4945]: I1014 15:35:48.199692 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 15:35:48 crc kubenswrapper[4945]: I1014 15:35:48.264807 4945 generic.go:334] "Generic (PLEG): container finished" podID="042f4fc3-10ec-450e-b3cf-77a11d71a8b7" containerID="463736e1b4ce1f16b5a23d2e524962c60720443527f70ae81106b5f41ac95227" exitCode=0 Oct 14 15:35:48 crc kubenswrapper[4945]: I1014 15:35:48.264903 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"042f4fc3-10ec-450e-b3cf-77a11d71a8b7","Type":"ContainerDied","Data":"463736e1b4ce1f16b5a23d2e524962c60720443527f70ae81106b5f41ac95227"} Oct 14 15:35:48 crc kubenswrapper[4945]: I1014 15:35:48.268393 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69","Type":"ContainerStarted","Data":"6fe89135579e241ee185769a1a7ceaaa98bb3c9d3bbc23909077785d73dd5fae"} Oct 14 15:35:48 crc kubenswrapper[4945]: I1014 15:35:48.268461 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a821c3a6-1324-4c0e-bcf4-1700cb3cfd69","Type":"ContainerStarted","Data":"10b586c7bb32ccd98e70b0854e10d866f5d6551fb259f0a8d4a023b986ba3300"} Oct 14 15:35:48 crc kubenswrapper[4945]: I1014 15:35:48.272192 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"524fdd1e-9a42-495f-a692-e4722ab3c4e2","Type":"ContainerStarted","Data":"3badb17ed7ebacdbe466dbbd3f9c2765a6194a879f387d6de5553071794a3830"} Oct 14 15:35:48 crc kubenswrapper[4945]: I1014 15:35:48.280901 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f22b1bbe-8060-4302-b929-567dc4f33173","Type":"ContainerStarted","Data":"cc3ccd6c550281a73c796cbf64a374b4394839751894510fb5a6e8f5e22e4818"} Oct 14 15:35:48 crc kubenswrapper[4945]: I1014 15:35:48.281340 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f22b1bbe-8060-4302-b929-567dc4f33173","Type":"ContainerStarted","Data":"351c7d2aca2d1b44a1537612900eed386bb4d3259b34ece5ec21703b6c84a728"} Oct 14 15:35:48 crc kubenswrapper[4945]: I1014 15:35:48.321996 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.321973353 podStartE2EDuration="3.321973353s" podCreationTimestamp="2025-10-14 15:35:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:35:48.310819934 +0000 UTC m=+1278.294868302" watchObservedRunningTime="2025-10-14 15:35:48.321973353 +0000 UTC m=+1278.306021731" Oct 14 15:35:48 crc kubenswrapper[4945]: I1014 15:35:48.337862 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.337843806 podStartE2EDuration="3.337843806s" podCreationTimestamp="2025-10-14 15:35:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:35:48.329406995 +0000 UTC m=+1278.313455383" watchObservedRunningTime="2025-10-14 15:35:48.337843806 +0000 UTC m=+1278.321892164" Oct 14 15:35:48 crc kubenswrapper[4945]: I1014 15:35:48.773139 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d65e47e7-2cb8-492c-aef0-6bc069d1f643" path="/var/lib/kubelet/pods/d65e47e7-2cb8-492c-aef0-6bc069d1f643/volumes" Oct 14 15:35:49 crc kubenswrapper[4945]: I1014 15:35:49.295507 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"524fdd1e-9a42-495f-a692-e4722ab3c4e2","Type":"ContainerStarted","Data":"1a8cc4cfb7c08a83e82f5ef551cb5102341405baf7dd9398e9be88e14e39f0e8"} Oct 14 15:35:49 crc kubenswrapper[4945]: I1014 15:35:49.299043 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"042f4fc3-10ec-450e-b3cf-77a11d71a8b7","Type":"ContainerStarted","Data":"f6ac1ec073aef015e90189cb146a47bc25e929e166ffd077a569aa210a2d8ea1"} Oct 14 15:35:49 crc kubenswrapper[4945]: I1014 15:35:49.299090 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"042f4fc3-10ec-450e-b3cf-77a11d71a8b7","Type":"ContainerStarted","Data":"2f1141fc6769331b2c6bab27432028a81bd9c02b34b9c4f3fc6bfd6490bb6ead"} Oct 14 15:35:49 crc kubenswrapper[4945]: I1014 15:35:49.328162 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.328134333 podStartE2EDuration="2.328134333s" podCreationTimestamp="2025-10-14 15:35:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:35:49.317157459 +0000 UTC m=+1279.301205827" watchObservedRunningTime="2025-10-14 15:35:49.328134333 +0000 UTC m=+1279.312182701" Oct 14 15:35:50 crc kubenswrapper[4945]: I1014 15:35:50.311689 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"042f4fc3-10ec-450e-b3cf-77a11d71a8b7","Type":"ContainerStarted","Data":"2a2bb480bc6a9cd6486184d8682d964362ae9b30331bfdf226fbba7f4e2802c3"} Oct 14 15:35:50 crc kubenswrapper[4945]: I1014 15:35:50.313102 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-conductor-0" Oct 14 15:35:50 crc kubenswrapper[4945]: I1014 15:35:50.313186 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-conductor-0" Oct 14 15:35:50 crc kubenswrapper[4945]: I1014 15:35:50.349106 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-conductor-0" podStartSLOduration=66.833778203 podStartE2EDuration="2m20.349089844s" podCreationTimestamp="2025-10-14 15:33:30 +0000 UTC" firstStartedPulling="2025-10-14 15:33:34.810335043 +0000 UTC m=+1144.794383411" lastFinishedPulling="2025-10-14 15:34:48.325646674 +0000 UTC m=+1218.309695052" observedRunningTime="2025-10-14 15:35:50.345696358 +0000 UTC m=+1280.329744736" watchObservedRunningTime="2025-10-14 15:35:50.349089844 +0000 UTC m=+1280.333138212" Oct 14 15:35:50 crc kubenswrapper[4945]: I1014 15:35:50.974131 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 15:35:50 crc kubenswrapper[4945]: I1014 15:35:50.974494 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 15:35:52 crc kubenswrapper[4945]: I1014 15:35:52.661736 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 15:35:55 crc kubenswrapper[4945]: I1014 15:35:55.961089 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 15:35:55 crc kubenswrapper[4945]: I1014 15:35:55.961530 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 15:35:55 crc kubenswrapper[4945]: I1014 15:35:55.974163 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 15:35:55 crc kubenswrapper[4945]: I1014 15:35:55.974227 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 15:35:56 crc kubenswrapper[4945]: I1014 15:35:56.985094 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a821c3a6-1324-4c0e-bcf4-1700cb3cfd69" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 15:35:56 crc kubenswrapper[4945]: I1014 15:35:56.985102 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a821c3a6-1324-4c0e-bcf4-1700cb3cfd69" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 15:35:56 crc kubenswrapper[4945]: I1014 15:35:56.999110 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f22b1bbe-8060-4302-b929-567dc4f33173" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.213:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 15:35:56 crc kubenswrapper[4945]: I1014 15:35:56.999171 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f22b1bbe-8060-4302-b929-567dc4f33173" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.213:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 15:35:57 crc kubenswrapper[4945]: I1014 15:35:57.661630 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 15:35:57 crc kubenswrapper[4945]: I1014 15:35:57.713687 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 15:35:58 crc kubenswrapper[4945]: I1014 15:35:58.462158 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 15:36:02 crc kubenswrapper[4945]: I1014 15:36:02.487272 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-conductor-0" Oct 14 15:36:02 crc kubenswrapper[4945]: I1014 15:36:02.488521 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-conductor-0" Oct 14 15:36:04 crc kubenswrapper[4945]: I1014 15:36:04.511978 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 15:36:05 crc kubenswrapper[4945]: I1014 15:36:05.967971 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 15:36:05 crc kubenswrapper[4945]: I1014 15:36:05.969901 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 15:36:05 crc kubenswrapper[4945]: I1014 15:36:05.970281 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 15:36:05 crc kubenswrapper[4945]: I1014 15:36:05.970452 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 15:36:05 crc kubenswrapper[4945]: I1014 15:36:05.977197 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 15:36:05 crc kubenswrapper[4945]: I1014 15:36:05.980312 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 15:36:05 crc kubenswrapper[4945]: I1014 15:36:05.981791 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 15:36:05 crc kubenswrapper[4945]: I1014 15:36:05.984068 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 15:36:05 crc kubenswrapper[4945]: I1014 15:36:05.997307 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 15:36:06 crc kubenswrapper[4945]: I1014 15:36:06.516638 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 15:36:14 crc kubenswrapper[4945]: I1014 15:36:14.602840 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 15:36:15 crc kubenswrapper[4945]: I1014 15:36:15.601246 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 15:36:18 crc kubenswrapper[4945]: I1014 15:36:18.744687 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="c7915803-bfbc-4150-854d-6039c685a771" containerName="rabbitmq" containerID="cri-o://18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753" gracePeriod=604796 Oct 14 15:36:19 crc kubenswrapper[4945]: I1014 15:36:19.517469 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="9642a3f8-eb9b-4f79-8448-602a5f7761d6" containerName="rabbitmq" containerID="cri-o://ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96" gracePeriod=604797 Oct 14 15:36:21 crc kubenswrapper[4945]: I1014 15:36:21.336463 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="c7915803-bfbc-4150-854d-6039c685a771" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 14 15:36:21 crc kubenswrapper[4945]: I1014 15:36:21.466489 4945 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9642a3f8-eb9b-4f79-8448-602a5f7761d6" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.310622 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.465080 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-config-data\") pod \"c7915803-bfbc-4150-854d-6039c685a771\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.465163 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-tls\") pod \"c7915803-bfbc-4150-854d-6039c685a771\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.465200 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-confd\") pod \"c7915803-bfbc-4150-854d-6039c685a771\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.465335 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2spth\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-kube-api-access-2spth\") pod \"c7915803-bfbc-4150-854d-6039c685a771\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.465398 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c7915803-bfbc-4150-854d-6039c685a771-pod-info\") pod \"c7915803-bfbc-4150-854d-6039c685a771\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.465427 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"c7915803-bfbc-4150-854d-6039c685a771\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.465463 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-plugins-conf\") pod \"c7915803-bfbc-4150-854d-6039c685a771\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.465542 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c7915803-bfbc-4150-854d-6039c685a771-erlang-cookie-secret\") pod \"c7915803-bfbc-4150-854d-6039c685a771\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.465577 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-erlang-cookie\") pod \"c7915803-bfbc-4150-854d-6039c685a771\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.465610 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-plugins\") pod \"c7915803-bfbc-4150-854d-6039c685a771\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.465638 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-server-conf\") pod \"c7915803-bfbc-4150-854d-6039c685a771\" (UID: \"c7915803-bfbc-4150-854d-6039c685a771\") " Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.466563 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c7915803-bfbc-4150-854d-6039c685a771" (UID: "c7915803-bfbc-4150-854d-6039c685a771"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.466724 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c7915803-bfbc-4150-854d-6039c685a771" (UID: "c7915803-bfbc-4150-854d-6039c685a771"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.466974 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c7915803-bfbc-4150-854d-6039c685a771" (UID: "c7915803-bfbc-4150-854d-6039c685a771"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.472270 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7915803-bfbc-4150-854d-6039c685a771-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c7915803-bfbc-4150-854d-6039c685a771" (UID: "c7915803-bfbc-4150-854d-6039c685a771"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.472311 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c7915803-bfbc-4150-854d-6039c685a771" (UID: "c7915803-bfbc-4150-854d-6039c685a771"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.472356 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c7915803-bfbc-4150-854d-6039c685a771-pod-info" (OuterVolumeSpecName: "pod-info") pod "c7915803-bfbc-4150-854d-6039c685a771" (UID: "c7915803-bfbc-4150-854d-6039c685a771"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.472387 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "c7915803-bfbc-4150-854d-6039c685a771" (UID: "c7915803-bfbc-4150-854d-6039c685a771"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.473592 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-kube-api-access-2spth" (OuterVolumeSpecName: "kube-api-access-2spth") pod "c7915803-bfbc-4150-854d-6039c685a771" (UID: "c7915803-bfbc-4150-854d-6039c685a771"). InnerVolumeSpecName "kube-api-access-2spth". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.525409 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-config-data" (OuterVolumeSpecName: "config-data") pod "c7915803-bfbc-4150-854d-6039c685a771" (UID: "c7915803-bfbc-4150-854d-6039c685a771"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.541819 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-server-conf" (OuterVolumeSpecName: "server-conf") pod "c7915803-bfbc-4150-854d-6039c685a771" (UID: "c7915803-bfbc-4150-854d-6039c685a771"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.568227 4945 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c7915803-bfbc-4150-854d-6039c685a771-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.568257 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.568270 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.568280 4945 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.568289 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.568297 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.568305 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2spth\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-kube-api-access-2spth\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.568313 4945 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c7915803-bfbc-4150-854d-6039c685a771-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.568340 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.568349 4945 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c7915803-bfbc-4150-854d-6039c685a771-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.589862 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.622949 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c7915803-bfbc-4150-854d-6039c685a771" (UID: "c7915803-bfbc-4150-854d-6039c685a771"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.669995 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c7915803-bfbc-4150-854d-6039c685a771-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.670069 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.707385 4945 generic.go:334] "Generic (PLEG): container finished" podID="c7915803-bfbc-4150-854d-6039c685a771" containerID="18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753" exitCode=0 Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.707449 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c7915803-bfbc-4150-854d-6039c685a771","Type":"ContainerDied","Data":"18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753"} Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.707488 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c7915803-bfbc-4150-854d-6039c685a771","Type":"ContainerDied","Data":"adea549cfc67ed39106c0fc0d2ed36f64d6a4b305ee21ef5341cb432e241c2ed"} Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.707524 4945 scope.go:117] "RemoveContainer" containerID="18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.707771 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.770355 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.774747 4945 scope.go:117] "RemoveContainer" containerID="9c60658f6099b1c229bfd7f9e639bfeb03e6915d14caaeb7e9f7520dc1e46c01" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.785732 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.796493 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 15:36:25 crc kubenswrapper[4945]: E1014 15:36:25.797175 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7915803-bfbc-4150-854d-6039c685a771" containerName="setup-container" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.797191 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7915803-bfbc-4150-854d-6039c685a771" containerName="setup-container" Oct 14 15:36:25 crc kubenswrapper[4945]: E1014 15:36:25.797222 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7915803-bfbc-4150-854d-6039c685a771" containerName="rabbitmq" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.797230 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7915803-bfbc-4150-854d-6039c685a771" containerName="rabbitmq" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.797396 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7915803-bfbc-4150-854d-6039c685a771" containerName="rabbitmq" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.798360 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.801948 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.802426 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.802917 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.803244 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-hgnp5" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.803434 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.803632 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.803814 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.813560 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.839735 4945 scope.go:117] "RemoveContainer" containerID="18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753" Oct 14 15:36:25 crc kubenswrapper[4945]: E1014 15:36:25.841459 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753\": container with ID starting with 18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753 not found: ID does not exist" containerID="18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.841499 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753"} err="failed to get container status \"18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753\": rpc error: code = NotFound desc = could not find container \"18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753\": container with ID starting with 18e337e44f65f4b23e16246993988328577eac125e7749a7a1fb96741f2d7753 not found: ID does not exist" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.841527 4945 scope.go:117] "RemoveContainer" containerID="9c60658f6099b1c229bfd7f9e639bfeb03e6915d14caaeb7e9f7520dc1e46c01" Oct 14 15:36:25 crc kubenswrapper[4945]: E1014 15:36:25.842728 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c60658f6099b1c229bfd7f9e639bfeb03e6915d14caaeb7e9f7520dc1e46c01\": container with ID starting with 9c60658f6099b1c229bfd7f9e639bfeb03e6915d14caaeb7e9f7520dc1e46c01 not found: ID does not exist" containerID="9c60658f6099b1c229bfd7f9e639bfeb03e6915d14caaeb7e9f7520dc1e46c01" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.842763 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c60658f6099b1c229bfd7f9e639bfeb03e6915d14caaeb7e9f7520dc1e46c01"} err="failed to get container status \"9c60658f6099b1c229bfd7f9e639bfeb03e6915d14caaeb7e9f7520dc1e46c01\": rpc error: code = NotFound desc = could not find container \"9c60658f6099b1c229bfd7f9e639bfeb03e6915d14caaeb7e9f7520dc1e46c01\": container with ID starting with 9c60658f6099b1c229bfd7f9e639bfeb03e6915d14caaeb7e9f7520dc1e46c01 not found: ID does not exist" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.875573 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/40cdc335-20ee-467f-bc15-4c5867d0fd2a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.875658 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/40cdc335-20ee-467f-bc15-4c5867d0fd2a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.875690 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pzf4\" (UniqueName: \"kubernetes.io/projected/40cdc335-20ee-467f-bc15-4c5867d0fd2a-kube-api-access-9pzf4\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.875865 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/40cdc335-20ee-467f-bc15-4c5867d0fd2a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.875991 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/40cdc335-20ee-467f-bc15-4c5867d0fd2a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.876124 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/40cdc335-20ee-467f-bc15-4c5867d0fd2a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.876157 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.876245 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40cdc335-20ee-467f-bc15-4c5867d0fd2a-config-data\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.876271 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/40cdc335-20ee-467f-bc15-4c5867d0fd2a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.876290 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/40cdc335-20ee-467f-bc15-4c5867d0fd2a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.876324 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/40cdc335-20ee-467f-bc15-4c5867d0fd2a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.979207 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/40cdc335-20ee-467f-bc15-4c5867d0fd2a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.979287 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.979367 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40cdc335-20ee-467f-bc15-4c5867d0fd2a-config-data\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.979398 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/40cdc335-20ee-467f-bc15-4c5867d0fd2a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.979425 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/40cdc335-20ee-467f-bc15-4c5867d0fd2a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.979450 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/40cdc335-20ee-467f-bc15-4c5867d0fd2a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.979539 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/40cdc335-20ee-467f-bc15-4c5867d0fd2a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.979579 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/40cdc335-20ee-467f-bc15-4c5867d0fd2a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.979603 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pzf4\" (UniqueName: \"kubernetes.io/projected/40cdc335-20ee-467f-bc15-4c5867d0fd2a-kube-api-access-9pzf4\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.979617 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.980706 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/40cdc335-20ee-467f-bc15-4c5867d0fd2a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.981001 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/40cdc335-20ee-467f-bc15-4c5867d0fd2a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.981138 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/40cdc335-20ee-467f-bc15-4c5867d0fd2a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.981217 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/40cdc335-20ee-467f-bc15-4c5867d0fd2a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.981717 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40cdc335-20ee-467f-bc15-4c5867d0fd2a-config-data\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.979636 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/40cdc335-20ee-467f-bc15-4c5867d0fd2a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:25 crc kubenswrapper[4945]: I1014 15:36:25.981804 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/40cdc335-20ee-467f-bc15-4c5867d0fd2a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.002205 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/40cdc335-20ee-467f-bc15-4c5867d0fd2a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.005227 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/40cdc335-20ee-467f-bc15-4c5867d0fd2a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.005506 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/40cdc335-20ee-467f-bc15-4c5867d0fd2a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.005959 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/40cdc335-20ee-467f-bc15-4c5867d0fd2a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.009201 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pzf4\" (UniqueName: \"kubernetes.io/projected/40cdc335-20ee-467f-bc15-4c5867d0fd2a-kube-api-access-9pzf4\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.038496 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"40cdc335-20ee-467f-bc15-4c5867d0fd2a\") " pod="openstack/rabbitmq-server-0" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.136568 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.144344 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.291317 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-confd\") pod \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.291412 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-server-conf\") pod \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.291510 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.291647 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9642a3f8-eb9b-4f79-8448-602a5f7761d6-pod-info\") pod \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.291687 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-config-data\") pod \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.291832 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-erlang-cookie\") pod \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.291993 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-tls\") pod \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.292075 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-plugins\") pod \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.292161 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9642a3f8-eb9b-4f79-8448-602a5f7761d6-erlang-cookie-secret\") pod \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.292211 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q2n2\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-kube-api-access-8q2n2\") pod \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.292322 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-plugins-conf\") pod \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\" (UID: \"9642a3f8-eb9b-4f79-8448-602a5f7761d6\") " Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.293556 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9642a3f8-eb9b-4f79-8448-602a5f7761d6" (UID: "9642a3f8-eb9b-4f79-8448-602a5f7761d6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.293798 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9642a3f8-eb9b-4f79-8448-602a5f7761d6" (UID: "9642a3f8-eb9b-4f79-8448-602a5f7761d6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.293976 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9642a3f8-eb9b-4f79-8448-602a5f7761d6" (UID: "9642a3f8-eb9b-4f79-8448-602a5f7761d6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.298696 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "9642a3f8-eb9b-4f79-8448-602a5f7761d6" (UID: "9642a3f8-eb9b-4f79-8448-602a5f7761d6"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.306707 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9642a3f8-eb9b-4f79-8448-602a5f7761d6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9642a3f8-eb9b-4f79-8448-602a5f7761d6" (UID: "9642a3f8-eb9b-4f79-8448-602a5f7761d6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.315486 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9642a3f8-eb9b-4f79-8448-602a5f7761d6-pod-info" (OuterVolumeSpecName: "pod-info") pod "9642a3f8-eb9b-4f79-8448-602a5f7761d6" (UID: "9642a3f8-eb9b-4f79-8448-602a5f7761d6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.315526 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9642a3f8-eb9b-4f79-8448-602a5f7761d6" (UID: "9642a3f8-eb9b-4f79-8448-602a5f7761d6"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.315652 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-kube-api-access-8q2n2" (OuterVolumeSpecName: "kube-api-access-8q2n2") pod "9642a3f8-eb9b-4f79-8448-602a5f7761d6" (UID: "9642a3f8-eb9b-4f79-8448-602a5f7761d6"). InnerVolumeSpecName "kube-api-access-8q2n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.353377 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-config-data" (OuterVolumeSpecName: "config-data") pod "9642a3f8-eb9b-4f79-8448-602a5f7761d6" (UID: "9642a3f8-eb9b-4f79-8448-602a5f7761d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.379796 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-server-conf" (OuterVolumeSpecName: "server-conf") pod "9642a3f8-eb9b-4f79-8448-602a5f7761d6" (UID: "9642a3f8-eb9b-4f79-8448-602a5f7761d6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.394635 4945 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.394677 4945 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.394689 4945 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9642a3f8-eb9b-4f79-8448-602a5f7761d6-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.394697 4945 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.394708 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.394718 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.394726 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.394735 4945 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9642a3f8-eb9b-4f79-8448-602a5f7761d6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.394743 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q2n2\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-kube-api-access-8q2n2\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.394751 4945 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9642a3f8-eb9b-4f79-8448-602a5f7761d6-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.425795 4945 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.436044 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9642a3f8-eb9b-4f79-8448-602a5f7761d6" (UID: "9642a3f8-eb9b-4f79-8448-602a5f7761d6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.497005 4945 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9642a3f8-eb9b-4f79-8448-602a5f7761d6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.497042 4945 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.661923 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.733386 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"40cdc335-20ee-467f-bc15-4c5867d0fd2a","Type":"ContainerStarted","Data":"e15bb55f954feb5f430118b75da94a7eeab7990ddb6d1a2bfc50e0fb3104184c"} Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.740454 4945 generic.go:334] "Generic (PLEG): container finished" podID="9642a3f8-eb9b-4f79-8448-602a5f7761d6" containerID="ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96" exitCode=0 Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.740510 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9642a3f8-eb9b-4f79-8448-602a5f7761d6","Type":"ContainerDied","Data":"ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96"} Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.740529 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.740982 4945 scope.go:117] "RemoveContainer" containerID="ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.743328 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9642a3f8-eb9b-4f79-8448-602a5f7761d6","Type":"ContainerDied","Data":"24d4b18d9bb1c46c9fffd46d7196f2130b4779a152df3350e5101c338988cf56"} Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.761135 4945 scope.go:117] "RemoveContainer" containerID="421c62871dd88c2e8493610a2bf3454c6bdaab0cc2d9530739b1bc7a8d4d5e01" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.775493 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7915803-bfbc-4150-854d-6039c685a771" path="/var/lib/kubelet/pods/c7915803-bfbc-4150-854d-6039c685a771/volumes" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.799007 4945 scope.go:117] "RemoveContainer" containerID="ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96" Oct 14 15:36:26 crc kubenswrapper[4945]: E1014 15:36:26.799332 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96\": container with ID starting with ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96 not found: ID does not exist" containerID="ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.799368 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96"} err="failed to get container status \"ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96\": rpc error: code = NotFound desc = could not find container \"ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96\": container with ID starting with ae47e5ad42002dd9b7ca134b1a772c72d39fe0153e60f2b3cf870acc973a7e96 not found: ID does not exist" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.799398 4945 scope.go:117] "RemoveContainer" containerID="421c62871dd88c2e8493610a2bf3454c6bdaab0cc2d9530739b1bc7a8d4d5e01" Oct 14 15:36:26 crc kubenswrapper[4945]: E1014 15:36:26.799634 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"421c62871dd88c2e8493610a2bf3454c6bdaab0cc2d9530739b1bc7a8d4d5e01\": container with ID starting with 421c62871dd88c2e8493610a2bf3454c6bdaab0cc2d9530739b1bc7a8d4d5e01 not found: ID does not exist" containerID="421c62871dd88c2e8493610a2bf3454c6bdaab0cc2d9530739b1bc7a8d4d5e01" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.799661 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"421c62871dd88c2e8493610a2bf3454c6bdaab0cc2d9530739b1bc7a8d4d5e01"} err="failed to get container status \"421c62871dd88c2e8493610a2bf3454c6bdaab0cc2d9530739b1bc7a8d4d5e01\": rpc error: code = NotFound desc = could not find container \"421c62871dd88c2e8493610a2bf3454c6bdaab0cc2d9530739b1bc7a8d4d5e01\": container with ID starting with 421c62871dd88c2e8493610a2bf3454c6bdaab0cc2d9530739b1bc7a8d4d5e01 not found: ID does not exist" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.803029 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.822513 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.836016 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 15:36:26 crc kubenswrapper[4945]: E1014 15:36:26.836406 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9642a3f8-eb9b-4f79-8448-602a5f7761d6" containerName="rabbitmq" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.836421 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9642a3f8-eb9b-4f79-8448-602a5f7761d6" containerName="rabbitmq" Oct 14 15:36:26 crc kubenswrapper[4945]: E1014 15:36:26.836459 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9642a3f8-eb9b-4f79-8448-602a5f7761d6" containerName="setup-container" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.836465 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="9642a3f8-eb9b-4f79-8448-602a5f7761d6" containerName="setup-container" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.836646 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="9642a3f8-eb9b-4f79-8448-602a5f7761d6" containerName="rabbitmq" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.837737 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.839902 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qpstp" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.840617 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.840639 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.840711 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.840854 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.841206 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.843837 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 14 15:36:26 crc kubenswrapper[4945]: I1014 15:36:26.852132 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.006757 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b52be853-0c59-48be-b4d0-3b2023310050-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.006796 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b52be853-0c59-48be-b4d0-3b2023310050-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.006857 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b52be853-0c59-48be-b4d0-3b2023310050-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.006950 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b52be853-0c59-48be-b4d0-3b2023310050-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.006990 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b52be853-0c59-48be-b4d0-3b2023310050-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.007037 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b52be853-0c59-48be-b4d0-3b2023310050-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.007061 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b52be853-0c59-48be-b4d0-3b2023310050-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.007082 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b52be853-0c59-48be-b4d0-3b2023310050-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.007105 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd6vr\" (UniqueName: \"kubernetes.io/projected/b52be853-0c59-48be-b4d0-3b2023310050-kube-api-access-qd6vr\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.007122 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.007160 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b52be853-0c59-48be-b4d0-3b2023310050-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.108845 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b52be853-0c59-48be-b4d0-3b2023310050-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.108910 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b52be853-0c59-48be-b4d0-3b2023310050-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.108936 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b52be853-0c59-48be-b4d0-3b2023310050-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.108953 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd6vr\" (UniqueName: \"kubernetes.io/projected/b52be853-0c59-48be-b4d0-3b2023310050-kube-api-access-qd6vr\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.108974 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.109011 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b52be853-0c59-48be-b4d0-3b2023310050-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.109060 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b52be853-0c59-48be-b4d0-3b2023310050-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.109075 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b52be853-0c59-48be-b4d0-3b2023310050-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.109106 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b52be853-0c59-48be-b4d0-3b2023310050-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.109158 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b52be853-0c59-48be-b4d0-3b2023310050-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.109183 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b52be853-0c59-48be-b4d0-3b2023310050-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.110024 4945 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.110179 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b52be853-0c59-48be-b4d0-3b2023310050-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.111316 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b52be853-0c59-48be-b4d0-3b2023310050-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.112724 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b52be853-0c59-48be-b4d0-3b2023310050-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.113086 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b52be853-0c59-48be-b4d0-3b2023310050-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.113958 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b52be853-0c59-48be-b4d0-3b2023310050-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.114615 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b52be853-0c59-48be-b4d0-3b2023310050-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.115305 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b52be853-0c59-48be-b4d0-3b2023310050-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.115618 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b52be853-0c59-48be-b4d0-3b2023310050-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.117009 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b52be853-0c59-48be-b4d0-3b2023310050-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.133785 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd6vr\" (UniqueName: \"kubernetes.io/projected/b52be853-0c59-48be-b4d0-3b2023310050-kube-api-access-qd6vr\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.138697 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b52be853-0c59-48be-b4d0-3b2023310050\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.157619 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.586215 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 15:36:27 crc kubenswrapper[4945]: I1014 15:36:27.753460 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b52be853-0c59-48be-b4d0-3b2023310050","Type":"ContainerStarted","Data":"b411ed6e56b5be90f05ccd7d3b3a70215b52ba8160eb903c13e0dbd0eeebc5f3"} Oct 14 15:36:28 crc kubenswrapper[4945]: I1014 15:36:28.778351 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9642a3f8-eb9b-4f79-8448-602a5f7761d6" path="/var/lib/kubelet/pods/9642a3f8-eb9b-4f79-8448-602a5f7761d6/volumes" Oct 14 15:36:28 crc kubenswrapper[4945]: I1014 15:36:28.779330 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"40cdc335-20ee-467f-bc15-4c5867d0fd2a","Type":"ContainerStarted","Data":"c2b44814f9d23ed6866562e8a1d521ca96b985f38a4687446fac04c43acb764f"} Oct 14 15:36:29 crc kubenswrapper[4945]: I1014 15:36:29.778832 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b52be853-0c59-48be-b4d0-3b2023310050","Type":"ContainerStarted","Data":"006fffc1d81112aa0b8a852bbc75e2d0b4645d78a017a1c47fe44d19a5e7f535"} Oct 14 15:37:01 crc kubenswrapper[4945]: I1014 15:37:01.126063 4945 generic.go:334] "Generic (PLEG): container finished" podID="40cdc335-20ee-467f-bc15-4c5867d0fd2a" containerID="c2b44814f9d23ed6866562e8a1d521ca96b985f38a4687446fac04c43acb764f" exitCode=0 Oct 14 15:37:01 crc kubenswrapper[4945]: I1014 15:37:01.126110 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"40cdc335-20ee-467f-bc15-4c5867d0fd2a","Type":"ContainerDied","Data":"c2b44814f9d23ed6866562e8a1d521ca96b985f38a4687446fac04c43acb764f"} Oct 14 15:37:02 crc kubenswrapper[4945]: I1014 15:37:02.137455 4945 generic.go:334] "Generic (PLEG): container finished" podID="b52be853-0c59-48be-b4d0-3b2023310050" containerID="006fffc1d81112aa0b8a852bbc75e2d0b4645d78a017a1c47fe44d19a5e7f535" exitCode=0 Oct 14 15:37:02 crc kubenswrapper[4945]: I1014 15:37:02.137531 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b52be853-0c59-48be-b4d0-3b2023310050","Type":"ContainerDied","Data":"006fffc1d81112aa0b8a852bbc75e2d0b4645d78a017a1c47fe44d19a5e7f535"} Oct 14 15:37:02 crc kubenswrapper[4945]: I1014 15:37:02.140895 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"40cdc335-20ee-467f-bc15-4c5867d0fd2a","Type":"ContainerStarted","Data":"7d3e77d0115a7df03233c87df781d5d78c78f42d6e2530ae74921eea8b9e6c7b"} Oct 14 15:37:02 crc kubenswrapper[4945]: I1014 15:37:02.141662 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 14 15:37:02 crc kubenswrapper[4945]: I1014 15:37:02.202270 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.202251771 podStartE2EDuration="37.202251771s" podCreationTimestamp="2025-10-14 15:36:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:37:02.195060284 +0000 UTC m=+1352.179108712" watchObservedRunningTime="2025-10-14 15:37:02.202251771 +0000 UTC m=+1352.186300139" Oct 14 15:37:03 crc kubenswrapper[4945]: I1014 15:37:03.150618 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b52be853-0c59-48be-b4d0-3b2023310050","Type":"ContainerStarted","Data":"8620aad1d11bc440b6204d16ed3ee9ea3c875451fe5f5d10ab0c1ea2634f1ae2"} Oct 14 15:37:03 crc kubenswrapper[4945]: I1014 15:37:03.151397 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:37:03 crc kubenswrapper[4945]: I1014 15:37:03.176852 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.176829189 podStartE2EDuration="37.176829189s" podCreationTimestamp="2025-10-14 15:36:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:37:03.171925188 +0000 UTC m=+1353.155973546" watchObservedRunningTime="2025-10-14 15:37:03.176829189 +0000 UTC m=+1353.160877557" Oct 14 15:37:16 crc kubenswrapper[4945]: I1014 15:37:16.140622 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 14 15:37:17 crc kubenswrapper[4945]: I1014 15:37:17.162117 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.563167 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zdrx6"] Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.566350 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.582871 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zdrx6"] Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.727656 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/994650de-b928-474c-914f-7414ac7243c2-catalog-content\") pod \"redhat-operators-zdrx6\" (UID: \"994650de-b928-474c-914f-7414ac7243c2\") " pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.727739 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/994650de-b928-474c-914f-7414ac7243c2-utilities\") pod \"redhat-operators-zdrx6\" (UID: \"994650de-b928-474c-914f-7414ac7243c2\") " pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.727793 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm6kk\" (UniqueName: \"kubernetes.io/projected/994650de-b928-474c-914f-7414ac7243c2-kube-api-access-fm6kk\") pod \"redhat-operators-zdrx6\" (UID: \"994650de-b928-474c-914f-7414ac7243c2\") " pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.795973 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.796037 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.829913 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm6kk\" (UniqueName: \"kubernetes.io/projected/994650de-b928-474c-914f-7414ac7243c2-kube-api-access-fm6kk\") pod \"redhat-operators-zdrx6\" (UID: \"994650de-b928-474c-914f-7414ac7243c2\") " pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.830418 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/994650de-b928-474c-914f-7414ac7243c2-catalog-content\") pod \"redhat-operators-zdrx6\" (UID: \"994650de-b928-474c-914f-7414ac7243c2\") " pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.830606 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/994650de-b928-474c-914f-7414ac7243c2-utilities\") pod \"redhat-operators-zdrx6\" (UID: \"994650de-b928-474c-914f-7414ac7243c2\") " pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.831029 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/994650de-b928-474c-914f-7414ac7243c2-catalog-content\") pod \"redhat-operators-zdrx6\" (UID: \"994650de-b928-474c-914f-7414ac7243c2\") " pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.831136 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/994650de-b928-474c-914f-7414ac7243c2-utilities\") pod \"redhat-operators-zdrx6\" (UID: \"994650de-b928-474c-914f-7414ac7243c2\") " pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.852233 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm6kk\" (UniqueName: \"kubernetes.io/projected/994650de-b928-474c-914f-7414ac7243c2-kube-api-access-fm6kk\") pod \"redhat-operators-zdrx6\" (UID: \"994650de-b928-474c-914f-7414ac7243c2\") " pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:16 crc kubenswrapper[4945]: I1014 15:38:16.901835 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:17 crc kubenswrapper[4945]: I1014 15:38:17.245361 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zdrx6"] Oct 14 15:38:17 crc kubenswrapper[4945]: I1014 15:38:17.998245 4945 generic.go:334] "Generic (PLEG): container finished" podID="994650de-b928-474c-914f-7414ac7243c2" containerID="6de9d01b7e3c8e0a1028262a269980cfc59266cd475a977a2ec939b639ae38d6" exitCode=0 Oct 14 15:38:17 crc kubenswrapper[4945]: I1014 15:38:17.998291 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdrx6" event={"ID":"994650de-b928-474c-914f-7414ac7243c2","Type":"ContainerDied","Data":"6de9d01b7e3c8e0a1028262a269980cfc59266cd475a977a2ec939b639ae38d6"} Oct 14 15:38:17 crc kubenswrapper[4945]: I1014 15:38:17.998315 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdrx6" event={"ID":"994650de-b928-474c-914f-7414ac7243c2","Type":"ContainerStarted","Data":"53d6ec6344f62b828348e95617f279b1edbf6e7fd57d055202a8efbf4392fc1c"} Oct 14 15:38:20 crc kubenswrapper[4945]: I1014 15:38:20.023535 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdrx6" event={"ID":"994650de-b928-474c-914f-7414ac7243c2","Type":"ContainerStarted","Data":"fcd613221ef6ec6d11caeda7a20c8fd5d78618b8569641e4dded55e7d2c2b3f7"} Oct 14 15:38:22 crc kubenswrapper[4945]: I1014 15:38:22.060640 4945 generic.go:334] "Generic (PLEG): container finished" podID="994650de-b928-474c-914f-7414ac7243c2" containerID="fcd613221ef6ec6d11caeda7a20c8fd5d78618b8569641e4dded55e7d2c2b3f7" exitCode=0 Oct 14 15:38:22 crc kubenswrapper[4945]: I1014 15:38:22.061117 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdrx6" event={"ID":"994650de-b928-474c-914f-7414ac7243c2","Type":"ContainerDied","Data":"fcd613221ef6ec6d11caeda7a20c8fd5d78618b8569641e4dded55e7d2c2b3f7"} Oct 14 15:38:23 crc kubenswrapper[4945]: I1014 15:38:23.073389 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdrx6" event={"ID":"994650de-b928-474c-914f-7414ac7243c2","Type":"ContainerStarted","Data":"b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f"} Oct 14 15:38:23 crc kubenswrapper[4945]: I1014 15:38:23.094630 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zdrx6" podStartSLOduration=2.273621811 podStartE2EDuration="7.094601422s" podCreationTimestamp="2025-10-14 15:38:16 +0000 UTC" firstStartedPulling="2025-10-14 15:38:18.002004596 +0000 UTC m=+1427.986052964" lastFinishedPulling="2025-10-14 15:38:22.822984207 +0000 UTC m=+1432.807032575" observedRunningTime="2025-10-14 15:38:23.092582994 +0000 UTC m=+1433.076631372" watchObservedRunningTime="2025-10-14 15:38:23.094601422 +0000 UTC m=+1433.078649800" Oct 14 15:38:26 crc kubenswrapper[4945]: I1014 15:38:26.902357 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:26 crc kubenswrapper[4945]: I1014 15:38:26.904007 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:27 crc kubenswrapper[4945]: I1014 15:38:27.955071 4945 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zdrx6" podUID="994650de-b928-474c-914f-7414ac7243c2" containerName="registry-server" probeResult="failure" output=< Oct 14 15:38:27 crc kubenswrapper[4945]: timeout: failed to connect service ":50051" within 1s Oct 14 15:38:27 crc kubenswrapper[4945]: > Oct 14 15:38:36 crc kubenswrapper[4945]: I1014 15:38:36.949915 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:36 crc kubenswrapper[4945]: I1014 15:38:36.994498 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:37 crc kubenswrapper[4945]: I1014 15:38:37.191398 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zdrx6"] Oct 14 15:38:38 crc kubenswrapper[4945]: I1014 15:38:38.205773 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zdrx6" podUID="994650de-b928-474c-914f-7414ac7243c2" containerName="registry-server" containerID="cri-o://b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f" gracePeriod=2 Oct 14 15:38:38 crc kubenswrapper[4945]: I1014 15:38:38.691230 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:38 crc kubenswrapper[4945]: I1014 15:38:38.862231 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/994650de-b928-474c-914f-7414ac7243c2-catalog-content\") pod \"994650de-b928-474c-914f-7414ac7243c2\" (UID: \"994650de-b928-474c-914f-7414ac7243c2\") " Oct 14 15:38:38 crc kubenswrapper[4945]: I1014 15:38:38.862367 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm6kk\" (UniqueName: \"kubernetes.io/projected/994650de-b928-474c-914f-7414ac7243c2-kube-api-access-fm6kk\") pod \"994650de-b928-474c-914f-7414ac7243c2\" (UID: \"994650de-b928-474c-914f-7414ac7243c2\") " Oct 14 15:38:38 crc kubenswrapper[4945]: I1014 15:38:38.862415 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/994650de-b928-474c-914f-7414ac7243c2-utilities\") pod \"994650de-b928-474c-914f-7414ac7243c2\" (UID: \"994650de-b928-474c-914f-7414ac7243c2\") " Oct 14 15:38:38 crc kubenswrapper[4945]: I1014 15:38:38.863186 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/994650de-b928-474c-914f-7414ac7243c2-utilities" (OuterVolumeSpecName: "utilities") pod "994650de-b928-474c-914f-7414ac7243c2" (UID: "994650de-b928-474c-914f-7414ac7243c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:38:38 crc kubenswrapper[4945]: I1014 15:38:38.878245 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/994650de-b928-474c-914f-7414ac7243c2-kube-api-access-fm6kk" (OuterVolumeSpecName: "kube-api-access-fm6kk") pod "994650de-b928-474c-914f-7414ac7243c2" (UID: "994650de-b928-474c-914f-7414ac7243c2"). InnerVolumeSpecName "kube-api-access-fm6kk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:38:38 crc kubenswrapper[4945]: I1014 15:38:38.944254 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/994650de-b928-474c-914f-7414ac7243c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "994650de-b928-474c-914f-7414ac7243c2" (UID: "994650de-b928-474c-914f-7414ac7243c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:38:38 crc kubenswrapper[4945]: I1014 15:38:38.964732 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/994650de-b928-474c-914f-7414ac7243c2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:38:38 crc kubenswrapper[4945]: I1014 15:38:38.964779 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm6kk\" (UniqueName: \"kubernetes.io/projected/994650de-b928-474c-914f-7414ac7243c2-kube-api-access-fm6kk\") on node \"crc\" DevicePath \"\"" Oct 14 15:38:38 crc kubenswrapper[4945]: I1014 15:38:38.964793 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/994650de-b928-474c-914f-7414ac7243c2-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.218309 4945 generic.go:334] "Generic (PLEG): container finished" podID="994650de-b928-474c-914f-7414ac7243c2" containerID="b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f" exitCode=0 Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.218364 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdrx6" event={"ID":"994650de-b928-474c-914f-7414ac7243c2","Type":"ContainerDied","Data":"b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f"} Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.218383 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zdrx6" Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.218407 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdrx6" event={"ID":"994650de-b928-474c-914f-7414ac7243c2","Type":"ContainerDied","Data":"53d6ec6344f62b828348e95617f279b1edbf6e7fd57d055202a8efbf4392fc1c"} Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.218434 4945 scope.go:117] "RemoveContainer" containerID="b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f" Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.251430 4945 scope.go:117] "RemoveContainer" containerID="fcd613221ef6ec6d11caeda7a20c8fd5d78618b8569641e4dded55e7d2c2b3f7" Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.263454 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zdrx6"] Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.274502 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zdrx6"] Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.287574 4945 scope.go:117] "RemoveContainer" containerID="6de9d01b7e3c8e0a1028262a269980cfc59266cd475a977a2ec939b639ae38d6" Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.321761 4945 scope.go:117] "RemoveContainer" containerID="b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f" Oct 14 15:38:39 crc kubenswrapper[4945]: E1014 15:38:39.322232 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f\": container with ID starting with b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f not found: ID does not exist" containerID="b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f" Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.322262 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f"} err="failed to get container status \"b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f\": rpc error: code = NotFound desc = could not find container \"b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f\": container with ID starting with b38a809981b6311770f005a993f74010b6a00fdb84d50b6ac11917a191521c7f not found: ID does not exist" Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.322281 4945 scope.go:117] "RemoveContainer" containerID="fcd613221ef6ec6d11caeda7a20c8fd5d78618b8569641e4dded55e7d2c2b3f7" Oct 14 15:38:39 crc kubenswrapper[4945]: E1014 15:38:39.322525 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcd613221ef6ec6d11caeda7a20c8fd5d78618b8569641e4dded55e7d2c2b3f7\": container with ID starting with fcd613221ef6ec6d11caeda7a20c8fd5d78618b8569641e4dded55e7d2c2b3f7 not found: ID does not exist" containerID="fcd613221ef6ec6d11caeda7a20c8fd5d78618b8569641e4dded55e7d2c2b3f7" Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.322546 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcd613221ef6ec6d11caeda7a20c8fd5d78618b8569641e4dded55e7d2c2b3f7"} err="failed to get container status \"fcd613221ef6ec6d11caeda7a20c8fd5d78618b8569641e4dded55e7d2c2b3f7\": rpc error: code = NotFound desc = could not find container \"fcd613221ef6ec6d11caeda7a20c8fd5d78618b8569641e4dded55e7d2c2b3f7\": container with ID starting with fcd613221ef6ec6d11caeda7a20c8fd5d78618b8569641e4dded55e7d2c2b3f7 not found: ID does not exist" Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.322560 4945 scope.go:117] "RemoveContainer" containerID="6de9d01b7e3c8e0a1028262a269980cfc59266cd475a977a2ec939b639ae38d6" Oct 14 15:38:39 crc kubenswrapper[4945]: E1014 15:38:39.322803 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6de9d01b7e3c8e0a1028262a269980cfc59266cd475a977a2ec939b639ae38d6\": container with ID starting with 6de9d01b7e3c8e0a1028262a269980cfc59266cd475a977a2ec939b639ae38d6 not found: ID does not exist" containerID="6de9d01b7e3c8e0a1028262a269980cfc59266cd475a977a2ec939b639ae38d6" Oct 14 15:38:39 crc kubenswrapper[4945]: I1014 15:38:39.322848 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6de9d01b7e3c8e0a1028262a269980cfc59266cd475a977a2ec939b639ae38d6"} err="failed to get container status \"6de9d01b7e3c8e0a1028262a269980cfc59266cd475a977a2ec939b639ae38d6\": rpc error: code = NotFound desc = could not find container \"6de9d01b7e3c8e0a1028262a269980cfc59266cd475a977a2ec939b639ae38d6\": container with ID starting with 6de9d01b7e3c8e0a1028262a269980cfc59266cd475a977a2ec939b639ae38d6 not found: ID does not exist" Oct 14 15:38:40 crc kubenswrapper[4945]: I1014 15:38:40.775492 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="994650de-b928-474c-914f-7414ac7243c2" path="/var/lib/kubelet/pods/994650de-b928-474c-914f-7414ac7243c2/volumes" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.605197 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sml7c"] Oct 14 15:38:41 crc kubenswrapper[4945]: E1014 15:38:41.605688 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="994650de-b928-474c-914f-7414ac7243c2" containerName="extract-content" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.605714 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="994650de-b928-474c-914f-7414ac7243c2" containerName="extract-content" Oct 14 15:38:41 crc kubenswrapper[4945]: E1014 15:38:41.605744 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="994650de-b928-474c-914f-7414ac7243c2" containerName="registry-server" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.605754 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="994650de-b928-474c-914f-7414ac7243c2" containerName="registry-server" Oct 14 15:38:41 crc kubenswrapper[4945]: E1014 15:38:41.605779 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="994650de-b928-474c-914f-7414ac7243c2" containerName="extract-utilities" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.605786 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="994650de-b928-474c-914f-7414ac7243c2" containerName="extract-utilities" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.605998 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="994650de-b928-474c-914f-7414ac7243c2" containerName="registry-server" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.607446 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.613548 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-utilities\") pod \"community-operators-sml7c\" (UID: \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\") " pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.613595 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-catalog-content\") pod \"community-operators-sml7c\" (UID: \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\") " pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.613632 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g7hn\" (UniqueName: \"kubernetes.io/projected/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-kube-api-access-8g7hn\") pod \"community-operators-sml7c\" (UID: \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\") " pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.624761 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sml7c"] Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.716643 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-utilities\") pod \"community-operators-sml7c\" (UID: \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\") " pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.716736 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-catalog-content\") pod \"community-operators-sml7c\" (UID: \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\") " pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.716824 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g7hn\" (UniqueName: \"kubernetes.io/projected/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-kube-api-access-8g7hn\") pod \"community-operators-sml7c\" (UID: \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\") " pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.717358 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-utilities\") pod \"community-operators-sml7c\" (UID: \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\") " pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.717501 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-catalog-content\") pod \"community-operators-sml7c\" (UID: \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\") " pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.737355 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g7hn\" (UniqueName: \"kubernetes.io/projected/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-kube-api-access-8g7hn\") pod \"community-operators-sml7c\" (UID: \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\") " pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:41 crc kubenswrapper[4945]: I1014 15:38:41.935497 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:42 crc kubenswrapper[4945]: I1014 15:38:42.413508 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sml7c"] Oct 14 15:38:43 crc kubenswrapper[4945]: I1014 15:38:43.008131 4945 scope.go:117] "RemoveContainer" containerID="a8af6b5fb71683b6fef75fe3d3a5f4f79023c41fa4e61dd955d4cb086d00d7db" Oct 14 15:38:43 crc kubenswrapper[4945]: I1014 15:38:43.037841 4945 scope.go:117] "RemoveContainer" containerID="d6ab6237fb57f889fb964006a306a22142b2a4c647f1fc836619036121255b99" Oct 14 15:38:43 crc kubenswrapper[4945]: I1014 15:38:43.059901 4945 scope.go:117] "RemoveContainer" containerID="960186b0407d5f8464bcdcc6228d6a3a3b0c2846418762cdec4bb21a8fb6e967" Oct 14 15:38:43 crc kubenswrapper[4945]: I1014 15:38:43.086235 4945 scope.go:117] "RemoveContainer" containerID="0ec9ff5df8f203d2f799b52eb146927423852e720b3bfe05be4c33e4a4b9a592" Oct 14 15:38:43 crc kubenswrapper[4945]: I1014 15:38:43.109305 4945 scope.go:117] "RemoveContainer" containerID="f7bb05eee912710a115a994a9fa6a23574b5055bc7048c2a5c0f429bc98b6492" Oct 14 15:38:43 crc kubenswrapper[4945]: I1014 15:38:43.274423 4945 generic.go:334] "Generic (PLEG): container finished" podID="72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" containerID="d554cabdc8d147a0410b2b4b37ac07c82d51790653ffd44a3077763670d9c37c" exitCode=0 Oct 14 15:38:43 crc kubenswrapper[4945]: I1014 15:38:43.274481 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml7c" event={"ID":"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764","Type":"ContainerDied","Data":"d554cabdc8d147a0410b2b4b37ac07c82d51790653ffd44a3077763670d9c37c"} Oct 14 15:38:43 crc kubenswrapper[4945]: I1014 15:38:43.274514 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml7c" event={"ID":"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764","Type":"ContainerStarted","Data":"df93acefd5cf2bb3618563a409e75373135f8f59da2c2f8d5ccdd3cd2bbe81ab"} Oct 14 15:38:44 crc kubenswrapper[4945]: I1014 15:38:44.286147 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml7c" event={"ID":"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764","Type":"ContainerStarted","Data":"8dfd6c7923c025bee4cc7cd91e08e8b74a3e4634286f64ab3268f6cb1ef2426b"} Oct 14 15:38:45 crc kubenswrapper[4945]: I1014 15:38:45.297900 4945 generic.go:334] "Generic (PLEG): container finished" podID="72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" containerID="8dfd6c7923c025bee4cc7cd91e08e8b74a3e4634286f64ab3268f6cb1ef2426b" exitCode=0 Oct 14 15:38:45 crc kubenswrapper[4945]: I1014 15:38:45.297988 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml7c" event={"ID":"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764","Type":"ContainerDied","Data":"8dfd6c7923c025bee4cc7cd91e08e8b74a3e4634286f64ab3268f6cb1ef2426b"} Oct 14 15:38:46 crc kubenswrapper[4945]: I1014 15:38:46.336999 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml7c" event={"ID":"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764","Type":"ContainerStarted","Data":"377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c"} Oct 14 15:38:46 crc kubenswrapper[4945]: I1014 15:38:46.359596 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sml7c" podStartSLOduration=2.948623696 podStartE2EDuration="5.359573414s" podCreationTimestamp="2025-10-14 15:38:41 +0000 UTC" firstStartedPulling="2025-10-14 15:38:43.2764303 +0000 UTC m=+1453.260478668" lastFinishedPulling="2025-10-14 15:38:45.687380018 +0000 UTC m=+1455.671428386" observedRunningTime="2025-10-14 15:38:46.35244492 +0000 UTC m=+1456.336493288" watchObservedRunningTime="2025-10-14 15:38:46.359573414 +0000 UTC m=+1456.343621782" Oct 14 15:38:46 crc kubenswrapper[4945]: I1014 15:38:46.796197 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:38:46 crc kubenswrapper[4945]: I1014 15:38:46.796490 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:38:51 crc kubenswrapper[4945]: I1014 15:38:51.936200 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:51 crc kubenswrapper[4945]: I1014 15:38:51.936924 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:52 crc kubenswrapper[4945]: I1014 15:38:52.002140 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:52 crc kubenswrapper[4945]: I1014 15:38:52.444612 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:52 crc kubenswrapper[4945]: I1014 15:38:52.501398 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sml7c"] Oct 14 15:38:54 crc kubenswrapper[4945]: I1014 15:38:54.412251 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sml7c" podUID="72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" containerName="registry-server" containerID="cri-o://377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c" gracePeriod=2 Oct 14 15:38:54 crc kubenswrapper[4945]: I1014 15:38:54.912374 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.014513 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-catalog-content\") pod \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\" (UID: \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\") " Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.014692 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g7hn\" (UniqueName: \"kubernetes.io/projected/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-kube-api-access-8g7hn\") pod \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\" (UID: \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\") " Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.014746 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-utilities\") pod \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\" (UID: \"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764\") " Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.016031 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-utilities" (OuterVolumeSpecName: "utilities") pod "72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" (UID: "72e7fb73-44f8-4cdb-baf8-4ed2a2e23764"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.022337 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-kube-api-access-8g7hn" (OuterVolumeSpecName: "kube-api-access-8g7hn") pod "72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" (UID: "72e7fb73-44f8-4cdb-baf8-4ed2a2e23764"). InnerVolumeSpecName "kube-api-access-8g7hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.079962 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" (UID: "72e7fb73-44f8-4cdb-baf8-4ed2a2e23764"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.116954 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g7hn\" (UniqueName: \"kubernetes.io/projected/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-kube-api-access-8g7hn\") on node \"crc\" DevicePath \"\"" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.117292 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.117306 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.424680 4945 generic.go:334] "Generic (PLEG): container finished" podID="72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" containerID="377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c" exitCode=0 Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.424733 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml7c" event={"ID":"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764","Type":"ContainerDied","Data":"377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c"} Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.424765 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml7c" event={"ID":"72e7fb73-44f8-4cdb-baf8-4ed2a2e23764","Type":"ContainerDied","Data":"df93acefd5cf2bb3618563a409e75373135f8f59da2c2f8d5ccdd3cd2bbe81ab"} Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.424773 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sml7c" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.424783 4945 scope.go:117] "RemoveContainer" containerID="377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.450649 4945 scope.go:117] "RemoveContainer" containerID="8dfd6c7923c025bee4cc7cd91e08e8b74a3e4634286f64ab3268f6cb1ef2426b" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.464275 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sml7c"] Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.473993 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sml7c"] Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.485973 4945 scope.go:117] "RemoveContainer" containerID="d554cabdc8d147a0410b2b4b37ac07c82d51790653ffd44a3077763670d9c37c" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.529365 4945 scope.go:117] "RemoveContainer" containerID="377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c" Oct 14 15:38:55 crc kubenswrapper[4945]: E1014 15:38:55.530383 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c\": container with ID starting with 377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c not found: ID does not exist" containerID="377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.530426 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c"} err="failed to get container status \"377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c\": rpc error: code = NotFound desc = could not find container \"377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c\": container with ID starting with 377f2a20930eee73a41177f3827c477250b385aca1f7594f4089541b6aacb59c not found: ID does not exist" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.530455 4945 scope.go:117] "RemoveContainer" containerID="8dfd6c7923c025bee4cc7cd91e08e8b74a3e4634286f64ab3268f6cb1ef2426b" Oct 14 15:38:55 crc kubenswrapper[4945]: E1014 15:38:55.530752 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dfd6c7923c025bee4cc7cd91e08e8b74a3e4634286f64ab3268f6cb1ef2426b\": container with ID starting with 8dfd6c7923c025bee4cc7cd91e08e8b74a3e4634286f64ab3268f6cb1ef2426b not found: ID does not exist" containerID="8dfd6c7923c025bee4cc7cd91e08e8b74a3e4634286f64ab3268f6cb1ef2426b" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.530780 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dfd6c7923c025bee4cc7cd91e08e8b74a3e4634286f64ab3268f6cb1ef2426b"} err="failed to get container status \"8dfd6c7923c025bee4cc7cd91e08e8b74a3e4634286f64ab3268f6cb1ef2426b\": rpc error: code = NotFound desc = could not find container \"8dfd6c7923c025bee4cc7cd91e08e8b74a3e4634286f64ab3268f6cb1ef2426b\": container with ID starting with 8dfd6c7923c025bee4cc7cd91e08e8b74a3e4634286f64ab3268f6cb1ef2426b not found: ID does not exist" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.530796 4945 scope.go:117] "RemoveContainer" containerID="d554cabdc8d147a0410b2b4b37ac07c82d51790653ffd44a3077763670d9c37c" Oct 14 15:38:55 crc kubenswrapper[4945]: E1014 15:38:55.531076 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d554cabdc8d147a0410b2b4b37ac07c82d51790653ffd44a3077763670d9c37c\": container with ID starting with d554cabdc8d147a0410b2b4b37ac07c82d51790653ffd44a3077763670d9c37c not found: ID does not exist" containerID="d554cabdc8d147a0410b2b4b37ac07c82d51790653ffd44a3077763670d9c37c" Oct 14 15:38:55 crc kubenswrapper[4945]: I1014 15:38:55.531114 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d554cabdc8d147a0410b2b4b37ac07c82d51790653ffd44a3077763670d9c37c"} err="failed to get container status \"d554cabdc8d147a0410b2b4b37ac07c82d51790653ffd44a3077763670d9c37c\": rpc error: code = NotFound desc = could not find container \"d554cabdc8d147a0410b2b4b37ac07c82d51790653ffd44a3077763670d9c37c\": container with ID starting with d554cabdc8d147a0410b2b4b37ac07c82d51790653ffd44a3077763670d9c37c not found: ID does not exist" Oct 14 15:38:56 crc kubenswrapper[4945]: I1014 15:38:56.790219 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" path="/var/lib/kubelet/pods/72e7fb73-44f8-4cdb-baf8-4ed2a2e23764/volumes" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.245950 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cbwct"] Oct 14 15:39:03 crc kubenswrapper[4945]: E1014 15:39:03.246701 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" containerName="extract-utilities" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.246713 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" containerName="extract-utilities" Oct 14 15:39:03 crc kubenswrapper[4945]: E1014 15:39:03.246750 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" containerName="extract-content" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.246756 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" containerName="extract-content" Oct 14 15:39:03 crc kubenswrapper[4945]: E1014 15:39:03.246769 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" containerName="registry-server" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.246777 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" containerName="registry-server" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.246990 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e7fb73-44f8-4cdb-baf8-4ed2a2e23764" containerName="registry-server" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.248273 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.264812 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbwct"] Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.391076 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a403eef-4a85-4567-896d-e9fc01d62626-utilities\") pod \"redhat-marketplace-cbwct\" (UID: \"5a403eef-4a85-4567-896d-e9fc01d62626\") " pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.391183 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqj5r\" (UniqueName: \"kubernetes.io/projected/5a403eef-4a85-4567-896d-e9fc01d62626-kube-api-access-lqj5r\") pod \"redhat-marketplace-cbwct\" (UID: \"5a403eef-4a85-4567-896d-e9fc01d62626\") " pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.391268 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a403eef-4a85-4567-896d-e9fc01d62626-catalog-content\") pod \"redhat-marketplace-cbwct\" (UID: \"5a403eef-4a85-4567-896d-e9fc01d62626\") " pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.492981 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a403eef-4a85-4567-896d-e9fc01d62626-utilities\") pod \"redhat-marketplace-cbwct\" (UID: \"5a403eef-4a85-4567-896d-e9fc01d62626\") " pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.493086 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqj5r\" (UniqueName: \"kubernetes.io/projected/5a403eef-4a85-4567-896d-e9fc01d62626-kube-api-access-lqj5r\") pod \"redhat-marketplace-cbwct\" (UID: \"5a403eef-4a85-4567-896d-e9fc01d62626\") " pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.493174 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a403eef-4a85-4567-896d-e9fc01d62626-catalog-content\") pod \"redhat-marketplace-cbwct\" (UID: \"5a403eef-4a85-4567-896d-e9fc01d62626\") " pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.493793 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a403eef-4a85-4567-896d-e9fc01d62626-catalog-content\") pod \"redhat-marketplace-cbwct\" (UID: \"5a403eef-4a85-4567-896d-e9fc01d62626\") " pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.494083 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a403eef-4a85-4567-896d-e9fc01d62626-utilities\") pod \"redhat-marketplace-cbwct\" (UID: \"5a403eef-4a85-4567-896d-e9fc01d62626\") " pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.527140 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqj5r\" (UniqueName: \"kubernetes.io/projected/5a403eef-4a85-4567-896d-e9fc01d62626-kube-api-access-lqj5r\") pod \"redhat-marketplace-cbwct\" (UID: \"5a403eef-4a85-4567-896d-e9fc01d62626\") " pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:03 crc kubenswrapper[4945]: I1014 15:39:03.566342 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:04 crc kubenswrapper[4945]: I1014 15:39:04.060901 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbwct"] Oct 14 15:39:04 crc kubenswrapper[4945]: I1014 15:39:04.514026 4945 generic.go:334] "Generic (PLEG): container finished" podID="5a403eef-4a85-4567-896d-e9fc01d62626" containerID="91218156b28410888f691d948db48e23cc49acb5fee70d068082cd283fbdb67f" exitCode=0 Oct 14 15:39:04 crc kubenswrapper[4945]: I1014 15:39:04.514313 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbwct" event={"ID":"5a403eef-4a85-4567-896d-e9fc01d62626","Type":"ContainerDied","Data":"91218156b28410888f691d948db48e23cc49acb5fee70d068082cd283fbdb67f"} Oct 14 15:39:04 crc kubenswrapper[4945]: I1014 15:39:04.514347 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbwct" event={"ID":"5a403eef-4a85-4567-896d-e9fc01d62626","Type":"ContainerStarted","Data":"e7dafc4f6db2db4e3d1f32e59a72d72bce8bd4cc1e7d0050d74449bfd273b52f"} Oct 14 15:39:06 crc kubenswrapper[4945]: I1014 15:39:06.542049 4945 generic.go:334] "Generic (PLEG): container finished" podID="5a403eef-4a85-4567-896d-e9fc01d62626" containerID="199c7d18dc6040dee5293661489d26e925e8cbcd565f3d5f8eeaa1933fab7091" exitCode=0 Oct 14 15:39:06 crc kubenswrapper[4945]: I1014 15:39:06.542089 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbwct" event={"ID":"5a403eef-4a85-4567-896d-e9fc01d62626","Type":"ContainerDied","Data":"199c7d18dc6040dee5293661489d26e925e8cbcd565f3d5f8eeaa1933fab7091"} Oct 14 15:39:07 crc kubenswrapper[4945]: I1014 15:39:07.553385 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbwct" event={"ID":"5a403eef-4a85-4567-896d-e9fc01d62626","Type":"ContainerStarted","Data":"75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293"} Oct 14 15:39:07 crc kubenswrapper[4945]: I1014 15:39:07.573470 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cbwct" podStartSLOduration=2.116415951 podStartE2EDuration="4.573454736s" podCreationTimestamp="2025-10-14 15:39:03 +0000 UTC" firstStartedPulling="2025-10-14 15:39:04.5164633 +0000 UTC m=+1474.500511678" lastFinishedPulling="2025-10-14 15:39:06.973502105 +0000 UTC m=+1476.957550463" observedRunningTime="2025-10-14 15:39:07.570280345 +0000 UTC m=+1477.554328713" watchObservedRunningTime="2025-10-14 15:39:07.573454736 +0000 UTC m=+1477.557503104" Oct 14 15:39:13 crc kubenswrapper[4945]: I1014 15:39:13.566613 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:13 crc kubenswrapper[4945]: I1014 15:39:13.567149 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:13 crc kubenswrapper[4945]: I1014 15:39:13.621321 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:13 crc kubenswrapper[4945]: I1014 15:39:13.689239 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:13 crc kubenswrapper[4945]: I1014 15:39:13.873775 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbwct"] Oct 14 15:39:15 crc kubenswrapper[4945]: I1014 15:39:15.630788 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cbwct" podUID="5a403eef-4a85-4567-896d-e9fc01d62626" containerName="registry-server" containerID="cri-o://75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293" gracePeriod=2 Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.110448 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.178057 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a403eef-4a85-4567-896d-e9fc01d62626-catalog-content\") pod \"5a403eef-4a85-4567-896d-e9fc01d62626\" (UID: \"5a403eef-4a85-4567-896d-e9fc01d62626\") " Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.178197 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a403eef-4a85-4567-896d-e9fc01d62626-utilities\") pod \"5a403eef-4a85-4567-896d-e9fc01d62626\" (UID: \"5a403eef-4a85-4567-896d-e9fc01d62626\") " Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.178234 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqj5r\" (UniqueName: \"kubernetes.io/projected/5a403eef-4a85-4567-896d-e9fc01d62626-kube-api-access-lqj5r\") pod \"5a403eef-4a85-4567-896d-e9fc01d62626\" (UID: \"5a403eef-4a85-4567-896d-e9fc01d62626\") " Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.180200 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a403eef-4a85-4567-896d-e9fc01d62626-utilities" (OuterVolumeSpecName: "utilities") pod "5a403eef-4a85-4567-896d-e9fc01d62626" (UID: "5a403eef-4a85-4567-896d-e9fc01d62626"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.184459 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a403eef-4a85-4567-896d-e9fc01d62626-kube-api-access-lqj5r" (OuterVolumeSpecName: "kube-api-access-lqj5r") pod "5a403eef-4a85-4567-896d-e9fc01d62626" (UID: "5a403eef-4a85-4567-896d-e9fc01d62626"). InnerVolumeSpecName "kube-api-access-lqj5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.192779 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a403eef-4a85-4567-896d-e9fc01d62626-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5a403eef-4a85-4567-896d-e9fc01d62626" (UID: "5a403eef-4a85-4567-896d-e9fc01d62626"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.280437 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a403eef-4a85-4567-896d-e9fc01d62626-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.280750 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a403eef-4a85-4567-896d-e9fc01d62626-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.280765 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqj5r\" (UniqueName: \"kubernetes.io/projected/5a403eef-4a85-4567-896d-e9fc01d62626-kube-api-access-lqj5r\") on node \"crc\" DevicePath \"\"" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.645062 4945 generic.go:334] "Generic (PLEG): container finished" podID="5a403eef-4a85-4567-896d-e9fc01d62626" containerID="75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293" exitCode=0 Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.645112 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbwct" event={"ID":"5a403eef-4a85-4567-896d-e9fc01d62626","Type":"ContainerDied","Data":"75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293"} Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.645146 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbwct" event={"ID":"5a403eef-4a85-4567-896d-e9fc01d62626","Type":"ContainerDied","Data":"e7dafc4f6db2db4e3d1f32e59a72d72bce8bd4cc1e7d0050d74449bfd273b52f"} Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.645168 4945 scope.go:117] "RemoveContainer" containerID="75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.645315 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbwct" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.680285 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbwct"] Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.681369 4945 scope.go:117] "RemoveContainer" containerID="199c7d18dc6040dee5293661489d26e925e8cbcd565f3d5f8eeaa1933fab7091" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.689051 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbwct"] Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.707162 4945 scope.go:117] "RemoveContainer" containerID="91218156b28410888f691d948db48e23cc49acb5fee70d068082cd283fbdb67f" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.758025 4945 scope.go:117] "RemoveContainer" containerID="75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293" Oct 14 15:39:16 crc kubenswrapper[4945]: E1014 15:39:16.758489 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293\": container with ID starting with 75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293 not found: ID does not exist" containerID="75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.758543 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293"} err="failed to get container status \"75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293\": rpc error: code = NotFound desc = could not find container \"75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293\": container with ID starting with 75d03574ae857ab2f148e5ff6b926a9e1f94c04d19fb483545372dca6e246293 not found: ID does not exist" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.758575 4945 scope.go:117] "RemoveContainer" containerID="199c7d18dc6040dee5293661489d26e925e8cbcd565f3d5f8eeaa1933fab7091" Oct 14 15:39:16 crc kubenswrapper[4945]: E1014 15:39:16.759007 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"199c7d18dc6040dee5293661489d26e925e8cbcd565f3d5f8eeaa1933fab7091\": container with ID starting with 199c7d18dc6040dee5293661489d26e925e8cbcd565f3d5f8eeaa1933fab7091 not found: ID does not exist" containerID="199c7d18dc6040dee5293661489d26e925e8cbcd565f3d5f8eeaa1933fab7091" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.759047 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"199c7d18dc6040dee5293661489d26e925e8cbcd565f3d5f8eeaa1933fab7091"} err="failed to get container status \"199c7d18dc6040dee5293661489d26e925e8cbcd565f3d5f8eeaa1933fab7091\": rpc error: code = NotFound desc = could not find container \"199c7d18dc6040dee5293661489d26e925e8cbcd565f3d5f8eeaa1933fab7091\": container with ID starting with 199c7d18dc6040dee5293661489d26e925e8cbcd565f3d5f8eeaa1933fab7091 not found: ID does not exist" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.759069 4945 scope.go:117] "RemoveContainer" containerID="91218156b28410888f691d948db48e23cc49acb5fee70d068082cd283fbdb67f" Oct 14 15:39:16 crc kubenswrapper[4945]: E1014 15:39:16.759340 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91218156b28410888f691d948db48e23cc49acb5fee70d068082cd283fbdb67f\": container with ID starting with 91218156b28410888f691d948db48e23cc49acb5fee70d068082cd283fbdb67f not found: ID does not exist" containerID="91218156b28410888f691d948db48e23cc49acb5fee70d068082cd283fbdb67f" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.759374 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91218156b28410888f691d948db48e23cc49acb5fee70d068082cd283fbdb67f"} err="failed to get container status \"91218156b28410888f691d948db48e23cc49acb5fee70d068082cd283fbdb67f\": rpc error: code = NotFound desc = could not find container \"91218156b28410888f691d948db48e23cc49acb5fee70d068082cd283fbdb67f\": container with ID starting with 91218156b28410888f691d948db48e23cc49acb5fee70d068082cd283fbdb67f not found: ID does not exist" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.774448 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a403eef-4a85-4567-896d-e9fc01d62626" path="/var/lib/kubelet/pods/5a403eef-4a85-4567-896d-e9fc01d62626/volumes" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.795494 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.795562 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.795613 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.796534 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8597fbae8d09b487635545f118989013ee775ca16d722567c5fc29358c69817b"} pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 15:39:16 crc kubenswrapper[4945]: I1014 15:39:16.796595 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" containerID="cri-o://8597fbae8d09b487635545f118989013ee775ca16d722567c5fc29358c69817b" gracePeriod=600 Oct 14 15:39:17 crc kubenswrapper[4945]: I1014 15:39:17.657551 4945 generic.go:334] "Generic (PLEG): container finished" podID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerID="8597fbae8d09b487635545f118989013ee775ca16d722567c5fc29358c69817b" exitCode=0 Oct 14 15:39:17 crc kubenswrapper[4945]: I1014 15:39:17.657615 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerDied","Data":"8597fbae8d09b487635545f118989013ee775ca16d722567c5fc29358c69817b"} Oct 14 15:39:17 crc kubenswrapper[4945]: I1014 15:39:17.658097 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerStarted","Data":"bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3"} Oct 14 15:39:17 crc kubenswrapper[4945]: I1014 15:39:17.658115 4945 scope.go:117] "RemoveContainer" containerID="4b33206c225bec396b0367980f7cae9b8b640b794421f9c820e87d051e7c6d3e" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.379713 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n6hrq"] Oct 14 15:39:20 crc kubenswrapper[4945]: E1014 15:39:20.381664 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a403eef-4a85-4567-896d-e9fc01d62626" containerName="registry-server" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.381744 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a403eef-4a85-4567-896d-e9fc01d62626" containerName="registry-server" Oct 14 15:39:20 crc kubenswrapper[4945]: E1014 15:39:20.381809 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a403eef-4a85-4567-896d-e9fc01d62626" containerName="extract-utilities" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.381893 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a403eef-4a85-4567-896d-e9fc01d62626" containerName="extract-utilities" Oct 14 15:39:20 crc kubenswrapper[4945]: E1014 15:39:20.382039 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a403eef-4a85-4567-896d-e9fc01d62626" containerName="extract-content" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.382091 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a403eef-4a85-4567-896d-e9fc01d62626" containerName="extract-content" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.382343 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a403eef-4a85-4567-896d-e9fc01d62626" containerName="registry-server" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.386339 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.407162 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n6hrq"] Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.458989 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-catalog-content\") pod \"certified-operators-n6hrq\" (UID: \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\") " pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.459082 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-utilities\") pod \"certified-operators-n6hrq\" (UID: \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\") " pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.459285 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvvgs\" (UniqueName: \"kubernetes.io/projected/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-kube-api-access-mvvgs\") pod \"certified-operators-n6hrq\" (UID: \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\") " pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.561115 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-catalog-content\") pod \"certified-operators-n6hrq\" (UID: \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\") " pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.561222 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-utilities\") pod \"certified-operators-n6hrq\" (UID: \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\") " pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.561301 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvvgs\" (UniqueName: \"kubernetes.io/projected/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-kube-api-access-mvvgs\") pod \"certified-operators-n6hrq\" (UID: \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\") " pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.561651 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-catalog-content\") pod \"certified-operators-n6hrq\" (UID: \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\") " pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.561751 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-utilities\") pod \"certified-operators-n6hrq\" (UID: \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\") " pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.585693 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvvgs\" (UniqueName: \"kubernetes.io/projected/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-kube-api-access-mvvgs\") pod \"certified-operators-n6hrq\" (UID: \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\") " pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:20 crc kubenswrapper[4945]: I1014 15:39:20.711236 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:21 crc kubenswrapper[4945]: I1014 15:39:21.219787 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n6hrq"] Oct 14 15:39:21 crc kubenswrapper[4945]: I1014 15:39:21.708430 4945 generic.go:334] "Generic (PLEG): container finished" podID="d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" containerID="ab4af72d8d2b61ecc4771a41c33531b967dc390a6630bfe716f9d428e899c5c8" exitCode=0 Oct 14 15:39:21 crc kubenswrapper[4945]: I1014 15:39:21.708496 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6hrq" event={"ID":"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5","Type":"ContainerDied","Data":"ab4af72d8d2b61ecc4771a41c33531b967dc390a6630bfe716f9d428e899c5c8"} Oct 14 15:39:21 crc kubenswrapper[4945]: I1014 15:39:21.710746 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6hrq" event={"ID":"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5","Type":"ContainerStarted","Data":"16e4d9464b5f2f21a45ae6eab27521632aa68f80612b6758afade8a51c2b906c"} Oct 14 15:39:23 crc kubenswrapper[4945]: I1014 15:39:23.756846 4945 generic.go:334] "Generic (PLEG): container finished" podID="d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" containerID="f51175774aa8f941a063f38525a335fbb26fdb78864430ecc4c24f84f5c8e211" exitCode=0 Oct 14 15:39:23 crc kubenswrapper[4945]: I1014 15:39:23.756942 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6hrq" event={"ID":"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5","Type":"ContainerDied","Data":"f51175774aa8f941a063f38525a335fbb26fdb78864430ecc4c24f84f5c8e211"} Oct 14 15:39:24 crc kubenswrapper[4945]: I1014 15:39:24.778636 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6hrq" event={"ID":"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5","Type":"ContainerStarted","Data":"53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c"} Oct 14 15:39:24 crc kubenswrapper[4945]: I1014 15:39:24.799491 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n6hrq" podStartSLOduration=2.288821762 podStartE2EDuration="4.799473332s" podCreationTimestamp="2025-10-14 15:39:20 +0000 UTC" firstStartedPulling="2025-10-14 15:39:21.710401659 +0000 UTC m=+1491.694450027" lastFinishedPulling="2025-10-14 15:39:24.221053229 +0000 UTC m=+1494.205101597" observedRunningTime="2025-10-14 15:39:24.795612192 +0000 UTC m=+1494.779660560" watchObservedRunningTime="2025-10-14 15:39:24.799473332 +0000 UTC m=+1494.783521690" Oct 14 15:39:30 crc kubenswrapper[4945]: I1014 15:39:30.711636 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:30 crc kubenswrapper[4945]: I1014 15:39:30.712274 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:30 crc kubenswrapper[4945]: I1014 15:39:30.779771 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:30 crc kubenswrapper[4945]: I1014 15:39:30.878049 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:31 crc kubenswrapper[4945]: I1014 15:39:31.017397 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n6hrq"] Oct 14 15:39:32 crc kubenswrapper[4945]: I1014 15:39:32.845786 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n6hrq" podUID="d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" containerName="registry-server" containerID="cri-o://53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c" gracePeriod=2 Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.291929 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.326944 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvvgs\" (UniqueName: \"kubernetes.io/projected/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-kube-api-access-mvvgs\") pod \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\" (UID: \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\") " Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.327019 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-catalog-content\") pod \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\" (UID: \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\") " Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.327071 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-utilities\") pod \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\" (UID: \"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5\") " Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.329454 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-utilities" (OuterVolumeSpecName: "utilities") pod "d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" (UID: "d9bc5f5d-9f3d-46ed-a019-a3977adc9da5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.335867 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-kube-api-access-mvvgs" (OuterVolumeSpecName: "kube-api-access-mvvgs") pod "d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" (UID: "d9bc5f5d-9f3d-46ed-a019-a3977adc9da5"). InnerVolumeSpecName "kube-api-access-mvvgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.412318 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" (UID: "d9bc5f5d-9f3d-46ed-a019-a3977adc9da5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.430307 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvvgs\" (UniqueName: \"kubernetes.io/projected/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-kube-api-access-mvvgs\") on node \"crc\" DevicePath \"\"" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.430351 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.430365 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.856775 4945 generic.go:334] "Generic (PLEG): container finished" podID="d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" containerID="53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c" exitCode=0 Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.856935 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6hrq" event={"ID":"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5","Type":"ContainerDied","Data":"53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c"} Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.857123 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6hrq" event={"ID":"d9bc5f5d-9f3d-46ed-a019-a3977adc9da5","Type":"ContainerDied","Data":"16e4d9464b5f2f21a45ae6eab27521632aa68f80612b6758afade8a51c2b906c"} Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.857148 4945 scope.go:117] "RemoveContainer" containerID="53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.857049 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6hrq" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.899658 4945 scope.go:117] "RemoveContainer" containerID="f51175774aa8f941a063f38525a335fbb26fdb78864430ecc4c24f84f5c8e211" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.902416 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n6hrq"] Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.911299 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n6hrq"] Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.943600 4945 scope.go:117] "RemoveContainer" containerID="ab4af72d8d2b61ecc4771a41c33531b967dc390a6630bfe716f9d428e899c5c8" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.985798 4945 scope.go:117] "RemoveContainer" containerID="53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c" Oct 14 15:39:33 crc kubenswrapper[4945]: E1014 15:39:33.986257 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c\": container with ID starting with 53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c not found: ID does not exist" containerID="53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.986300 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c"} err="failed to get container status \"53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c\": rpc error: code = NotFound desc = could not find container \"53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c\": container with ID starting with 53ea43e8457157da5a941cd21d5fd1cae45dbe0c9d34954a00e109d76bf82a2c not found: ID does not exist" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.986332 4945 scope.go:117] "RemoveContainer" containerID="f51175774aa8f941a063f38525a335fbb26fdb78864430ecc4c24f84f5c8e211" Oct 14 15:39:33 crc kubenswrapper[4945]: E1014 15:39:33.986735 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f51175774aa8f941a063f38525a335fbb26fdb78864430ecc4c24f84f5c8e211\": container with ID starting with f51175774aa8f941a063f38525a335fbb26fdb78864430ecc4c24f84f5c8e211 not found: ID does not exist" containerID="f51175774aa8f941a063f38525a335fbb26fdb78864430ecc4c24f84f5c8e211" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.986796 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f51175774aa8f941a063f38525a335fbb26fdb78864430ecc4c24f84f5c8e211"} err="failed to get container status \"f51175774aa8f941a063f38525a335fbb26fdb78864430ecc4c24f84f5c8e211\": rpc error: code = NotFound desc = could not find container \"f51175774aa8f941a063f38525a335fbb26fdb78864430ecc4c24f84f5c8e211\": container with ID starting with f51175774aa8f941a063f38525a335fbb26fdb78864430ecc4c24f84f5c8e211 not found: ID does not exist" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.986832 4945 scope.go:117] "RemoveContainer" containerID="ab4af72d8d2b61ecc4771a41c33531b967dc390a6630bfe716f9d428e899c5c8" Oct 14 15:39:33 crc kubenswrapper[4945]: E1014 15:39:33.987185 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab4af72d8d2b61ecc4771a41c33531b967dc390a6630bfe716f9d428e899c5c8\": container with ID starting with ab4af72d8d2b61ecc4771a41c33531b967dc390a6630bfe716f9d428e899c5c8 not found: ID does not exist" containerID="ab4af72d8d2b61ecc4771a41c33531b967dc390a6630bfe716f9d428e899c5c8" Oct 14 15:39:33 crc kubenswrapper[4945]: I1014 15:39:33.987216 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab4af72d8d2b61ecc4771a41c33531b967dc390a6630bfe716f9d428e899c5c8"} err="failed to get container status \"ab4af72d8d2b61ecc4771a41c33531b967dc390a6630bfe716f9d428e899c5c8\": rpc error: code = NotFound desc = could not find container \"ab4af72d8d2b61ecc4771a41c33531b967dc390a6630bfe716f9d428e899c5c8\": container with ID starting with ab4af72d8d2b61ecc4771a41c33531b967dc390a6630bfe716f9d428e899c5c8 not found: ID does not exist" Oct 14 15:39:34 crc kubenswrapper[4945]: I1014 15:39:34.774528 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" path="/var/lib/kubelet/pods/d9bc5f5d-9f3d-46ed-a019-a3977adc9da5/volumes" Oct 14 15:39:43 crc kubenswrapper[4945]: I1014 15:39:43.240988 4945 scope.go:117] "RemoveContainer" containerID="b31e359b3fa23de12774859bfd90254efd89dace829bc5094b2504daa003f084" Oct 14 15:39:43 crc kubenswrapper[4945]: I1014 15:39:43.301947 4945 scope.go:117] "RemoveContainer" containerID="0e69d3d567d012f6328607ddd7ea353000e2d899addbff317eed50910ffc2ad6" Oct 14 15:40:43 crc kubenswrapper[4945]: I1014 15:40:43.450210 4945 scope.go:117] "RemoveContainer" containerID="766f810296ec7a5a93ef1b347b3515dc99539fc1028d898e008f6876c2c041ca" Oct 14 15:40:43 crc kubenswrapper[4945]: I1014 15:40:43.484801 4945 scope.go:117] "RemoveContainer" containerID="4da9d23841a555fd547a272e8fc433ccbaa0dd14c62e71ff5b0495896e9af169" Oct 14 15:40:43 crc kubenswrapper[4945]: I1014 15:40:43.513837 4945 scope.go:117] "RemoveContainer" containerID="ce4a3c3cedfd690774e20edda998c949e732f95fa275b6d433dd83fee61c0f76" Oct 14 15:40:43 crc kubenswrapper[4945]: I1014 15:40:43.539970 4945 scope.go:117] "RemoveContainer" containerID="e2fc55e977d5b6625355202d173ead6e7f1899f58d98552f0ccc6b6f92e1bea9" Oct 14 15:40:43 crc kubenswrapper[4945]: I1014 15:40:43.560175 4945 scope.go:117] "RemoveContainer" containerID="dcc670df4f7e3d081db51c05697440d42d5a6f340a7b34d370428e2d0c7dd19c" Oct 14 15:40:43 crc kubenswrapper[4945]: I1014 15:40:43.582586 4945 scope.go:117] "RemoveContainer" containerID="efbbf41db8b3ea1be2d941267bb61587b08fe52e37bf4ff26d840f59dc1cedad" Oct 14 15:41:18 crc kubenswrapper[4945]: I1014 15:41:18.041203 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-m6rmj"] Oct 14 15:41:18 crc kubenswrapper[4945]: I1014 15:41:18.051120 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-m6rmj"] Oct 14 15:41:18 crc kubenswrapper[4945]: I1014 15:41:18.061070 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-hmwc5"] Oct 14 15:41:18 crc kubenswrapper[4945]: I1014 15:41:18.068403 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-hmwc5"] Oct 14 15:41:18 crc kubenswrapper[4945]: I1014 15:41:18.777248 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a886bcc8-a794-45d1-a4f4-90ffabf9fb58" path="/var/lib/kubelet/pods/a886bcc8-a794-45d1-a4f4-90ffabf9fb58/volumes" Oct 14 15:41:18 crc kubenswrapper[4945]: I1014 15:41:18.778219 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c47e35eb-9ce0-4ce8-9b41-defa17743ccc" path="/var/lib/kubelet/pods/c47e35eb-9ce0-4ce8-9b41-defa17743ccc/volumes" Oct 14 15:41:23 crc kubenswrapper[4945]: I1014 15:41:23.040131 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-kknvq"] Oct 14 15:41:23 crc kubenswrapper[4945]: I1014 15:41:23.047022 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-kknvq"] Oct 14 15:41:24 crc kubenswrapper[4945]: I1014 15:41:24.778425 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf94dd02-5151-404b-8444-24a1cb34ca64" path="/var/lib/kubelet/pods/cf94dd02-5151-404b-8444-24a1cb34ca64/volumes" Oct 14 15:41:28 crc kubenswrapper[4945]: I1014 15:41:28.027450 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-812a-account-create-cf984"] Oct 14 15:41:28 crc kubenswrapper[4945]: I1014 15:41:28.036013 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ab43-account-create-sksdh"] Oct 14 15:41:28 crc kubenswrapper[4945]: I1014 15:41:28.047123 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-812a-account-create-cf984"] Oct 14 15:41:28 crc kubenswrapper[4945]: I1014 15:41:28.058471 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-ab43-account-create-sksdh"] Oct 14 15:41:28 crc kubenswrapper[4945]: I1014 15:41:28.774138 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51f09adf-1049-4876-b658-4552d03dc683" path="/var/lib/kubelet/pods/51f09adf-1049-4876-b658-4552d03dc683/volumes" Oct 14 15:41:28 crc kubenswrapper[4945]: I1014 15:41:28.775159 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e394f9f-f797-4659-b933-432c563567d8" path="/var/lib/kubelet/pods/7e394f9f-f797-4659-b933-432c563567d8/volumes" Oct 14 15:41:33 crc kubenswrapper[4945]: I1014 15:41:33.029852 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-534f-account-create-nl9hz"] Oct 14 15:41:33 crc kubenswrapper[4945]: I1014 15:41:33.038858 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-534f-account-create-nl9hz"] Oct 14 15:41:34 crc kubenswrapper[4945]: I1014 15:41:34.782087 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eed35ce9-830a-4746-a5f7-e2e7cfd44d75" path="/var/lib/kubelet/pods/eed35ce9-830a-4746-a5f7-e2e7cfd44d75/volumes" Oct 14 15:41:35 crc kubenswrapper[4945]: I1014 15:41:35.040214 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-9k955"] Oct 14 15:41:35 crc kubenswrapper[4945]: I1014 15:41:35.054196 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-9k955"] Oct 14 15:41:36 crc kubenswrapper[4945]: I1014 15:41:36.039054 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-7cz6h"] Oct 14 15:41:36 crc kubenswrapper[4945]: I1014 15:41:36.049020 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-tfzfz"] Oct 14 15:41:36 crc kubenswrapper[4945]: I1014 15:41:36.064439 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-7cz6h"] Oct 14 15:41:36 crc kubenswrapper[4945]: I1014 15:41:36.072201 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-tfzfz"] Oct 14 15:41:36 crc kubenswrapper[4945]: I1014 15:41:36.775036 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c9d06f9-312e-464b-9438-4f2f34a4daf0" path="/var/lib/kubelet/pods/1c9d06f9-312e-464b-9438-4f2f34a4daf0/volumes" Oct 14 15:41:36 crc kubenswrapper[4945]: I1014 15:41:36.776047 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="826c5051-a7b9-4f9a-98dd-cb1954e00a10" path="/var/lib/kubelet/pods/826c5051-a7b9-4f9a-98dd-cb1954e00a10/volumes" Oct 14 15:41:36 crc kubenswrapper[4945]: I1014 15:41:36.776677 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="887c4eba-a780-4417-a255-528cfd654d8b" path="/var/lib/kubelet/pods/887c4eba-a780-4417-a255-528cfd654d8b/volumes" Oct 14 15:41:43 crc kubenswrapper[4945]: I1014 15:41:43.686158 4945 scope.go:117] "RemoveContainer" containerID="4824edc86a1d31578b95250e8420fb346b0eea01d539f2b095e8a2e673d07b5c" Oct 14 15:41:43 crc kubenswrapper[4945]: I1014 15:41:43.714974 4945 scope.go:117] "RemoveContainer" containerID="906273dd7ab06c572849d59c59594ae665e8df032223b846c25cd5fe75e2ab8b" Oct 14 15:41:43 crc kubenswrapper[4945]: I1014 15:41:43.792520 4945 scope.go:117] "RemoveContainer" containerID="32c05eacfa9988f89fc32fd69945888db53d27247a1244a098b0ce8e2aaaaf43" Oct 14 15:41:43 crc kubenswrapper[4945]: I1014 15:41:43.831062 4945 scope.go:117] "RemoveContainer" containerID="29f039448f9d5fab5df0aa15d63401bfe7a6c6bb55809f04f1e34935e25835ce" Oct 14 15:41:43 crc kubenswrapper[4945]: I1014 15:41:43.871686 4945 scope.go:117] "RemoveContainer" containerID="61af3dd7ce199caf3ff5f9ca2275d347c15460ce7d65f14b36456335f3de13ed" Oct 14 15:41:43 crc kubenswrapper[4945]: I1014 15:41:43.938626 4945 scope.go:117] "RemoveContainer" containerID="887307eb130aeee4a121c651b4244748e960cb1684871a6e58db2aa322bcec1c" Oct 14 15:41:43 crc kubenswrapper[4945]: I1014 15:41:43.971097 4945 scope.go:117] "RemoveContainer" containerID="cb47933954470f605e8b5db33fc6d09991dd6f92a7fd05fa837dee287180cba3" Oct 14 15:41:43 crc kubenswrapper[4945]: I1014 15:41:43.991471 4945 scope.go:117] "RemoveContainer" containerID="8b09de0a388a4af4b207025b85d07804726d41e9b3a99c8f0e1578134df48dc1" Oct 14 15:41:44 crc kubenswrapper[4945]: I1014 15:41:44.014361 4945 scope.go:117] "RemoveContainer" containerID="5df90a957d6f8c9f7a3786d257950ae5fd5cf6972ecd9a1dc2a77b1fa4847c1a" Oct 14 15:41:45 crc kubenswrapper[4945]: I1014 15:41:45.037830 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b8b8-account-create-fd6fj"] Oct 14 15:41:45 crc kubenswrapper[4945]: I1014 15:41:45.050794 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-dc8c-account-create-k2465"] Oct 14 15:41:45 crc kubenswrapper[4945]: I1014 15:41:45.061831 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-dc8c-account-create-k2465"] Oct 14 15:41:45 crc kubenswrapper[4945]: I1014 15:41:45.069891 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b8b8-account-create-fd6fj"] Oct 14 15:41:46 crc kubenswrapper[4945]: I1014 15:41:46.039300 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-01ac-account-create-c6467"] Oct 14 15:41:46 crc kubenswrapper[4945]: I1014 15:41:46.071644 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-d6w8h"] Oct 14 15:41:46 crc kubenswrapper[4945]: I1014 15:41:46.088184 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-01ac-account-create-c6467"] Oct 14 15:41:46 crc kubenswrapper[4945]: I1014 15:41:46.095783 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-d6w8h"] Oct 14 15:41:46 crc kubenswrapper[4945]: I1014 15:41:46.775793 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c4ea6c9-6e27-4af8-8115-7582222b971f" path="/var/lib/kubelet/pods/0c4ea6c9-6e27-4af8-8115-7582222b971f/volumes" Oct 14 15:41:46 crc kubenswrapper[4945]: I1014 15:41:46.776331 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="155850d5-d5d8-419e-8bfb-580ae259b00c" path="/var/lib/kubelet/pods/155850d5-d5d8-419e-8bfb-580ae259b00c/volumes" Oct 14 15:41:46 crc kubenswrapper[4945]: I1014 15:41:46.777002 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c11ab49-aef7-4d60-b287-147be1668631" path="/var/lib/kubelet/pods/1c11ab49-aef7-4d60-b287-147be1668631/volumes" Oct 14 15:41:46 crc kubenswrapper[4945]: I1014 15:41:46.777484 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ebee178-706b-44af-9df7-d7292d028a28" path="/var/lib/kubelet/pods/3ebee178-706b-44af-9df7-d7292d028a28/volumes" Oct 14 15:41:46 crc kubenswrapper[4945]: I1014 15:41:46.795761 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:41:46 crc kubenswrapper[4945]: I1014 15:41:46.795816 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:42:06 crc kubenswrapper[4945]: I1014 15:42:06.043104 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-db-create-l5q2x"] Oct 14 15:42:06 crc kubenswrapper[4945]: I1014 15:42:06.053200 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-db-create-l5q2x"] Oct 14 15:42:06 crc kubenswrapper[4945]: I1014 15:42:06.776656 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56ce1365-283f-4f53-8563-60efbce145f7" path="/var/lib/kubelet/pods/56ce1365-283f-4f53-8563-60efbce145f7/volumes" Oct 14 15:42:16 crc kubenswrapper[4945]: I1014 15:42:16.795914 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:42:16 crc kubenswrapper[4945]: I1014 15:42:16.797265 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:42:25 crc kubenswrapper[4945]: I1014 15:42:25.036161 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-204d-account-create-v7l8c"] Oct 14 15:42:25 crc kubenswrapper[4945]: I1014 15:42:25.046949 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-204d-account-create-v7l8c"] Oct 14 15:42:26 crc kubenswrapper[4945]: I1014 15:42:26.773954 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6a0ed65-3b9d-460f-9493-4c1280e1b6cc" path="/var/lib/kubelet/pods/e6a0ed65-3b9d-460f-9493-4c1280e1b6cc/volumes" Oct 14 15:42:39 crc kubenswrapper[4945]: I1014 15:42:39.033903 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-mf7rz"] Oct 14 15:42:39 crc kubenswrapper[4945]: I1014 15:42:39.043159 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-mf7rz"] Oct 14 15:42:39 crc kubenswrapper[4945]: I1014 15:42:39.053013 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-kjrhb"] Oct 14 15:42:39 crc kubenswrapper[4945]: I1014 15:42:39.062900 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-kjrhb"] Oct 14 15:42:40 crc kubenswrapper[4945]: I1014 15:42:40.775439 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13783d41-0f7f-499e-bfbe-eaa0c6440d6e" path="/var/lib/kubelet/pods/13783d41-0f7f-499e-bfbe-eaa0c6440d6e/volumes" Oct 14 15:42:40 crc kubenswrapper[4945]: I1014 15:42:40.776013 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbc6053c-32ba-4e9f-86b7-8590cff47464" path="/var/lib/kubelet/pods/bbc6053c-32ba-4e9f-86b7-8590cff47464/volumes" Oct 14 15:42:44 crc kubenswrapper[4945]: I1014 15:42:44.267326 4945 scope.go:117] "RemoveContainer" containerID="1337f9899efdf21e01d543d6446b2ee1c8e1c275cefd765b667d354f0ad1991c" Oct 14 15:42:44 crc kubenswrapper[4945]: I1014 15:42:44.299120 4945 scope.go:117] "RemoveContainer" containerID="8b31e8950fa06614de88fdc5634115683e7ff58dddda2229d1dc4d96c4127b22" Oct 14 15:42:44 crc kubenswrapper[4945]: I1014 15:42:44.375623 4945 scope.go:117] "RemoveContainer" containerID="26f8c2f34be669431a325b5a51793e9db83dce121f3d64c1f0c493e5f23c7135" Oct 14 15:42:44 crc kubenswrapper[4945]: I1014 15:42:44.411755 4945 scope.go:117] "RemoveContainer" containerID="a70321bd680748812f1478a4e9f2be939302f7d6086028defe66d0c8b4d7c84f" Oct 14 15:42:44 crc kubenswrapper[4945]: I1014 15:42:44.482230 4945 scope.go:117] "RemoveContainer" containerID="760a37312437f03b6331c881c11c4ef944fb136c3556d5ba0f7bd495b2b68dc7" Oct 14 15:42:44 crc kubenswrapper[4945]: I1014 15:42:44.518141 4945 scope.go:117] "RemoveContainer" containerID="7eea8c21fbbf3dd6bcf0bc3b3a1bc387f6d0b95acd41720addf787b8f431ea1d" Oct 14 15:42:44 crc kubenswrapper[4945]: I1014 15:42:44.561187 4945 scope.go:117] "RemoveContainer" containerID="2a8e9691658fc574d29b3377c2db31609ea425cd32d51003ba215c90ed4e78c1" Oct 14 15:42:44 crc kubenswrapper[4945]: I1014 15:42:44.587265 4945 scope.go:117] "RemoveContainer" containerID="9ea6c19084af198c8763b3c730d8213fd1fbaeb122a274c1c4cf1b805545399d" Oct 14 15:42:46 crc kubenswrapper[4945]: I1014 15:42:46.795667 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:42:46 crc kubenswrapper[4945]: I1014 15:42:46.795910 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:42:46 crc kubenswrapper[4945]: I1014 15:42:46.795948 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:42:46 crc kubenswrapper[4945]: I1014 15:42:46.796445 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3"} pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 15:42:46 crc kubenswrapper[4945]: I1014 15:42:46.796494 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" containerID="cri-o://bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" gracePeriod=600 Oct 14 15:42:46 crc kubenswrapper[4945]: E1014 15:42:46.915407 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:42:47 crc kubenswrapper[4945]: I1014 15:42:47.711766 4945 generic.go:334] "Generic (PLEG): container finished" podID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" exitCode=0 Oct 14 15:42:47 crc kubenswrapper[4945]: I1014 15:42:47.711832 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerDied","Data":"bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3"} Oct 14 15:42:47 crc kubenswrapper[4945]: I1014 15:42:47.712257 4945 scope.go:117] "RemoveContainer" containerID="8597fbae8d09b487635545f118989013ee775ca16d722567c5fc29358c69817b" Oct 14 15:42:47 crc kubenswrapper[4945]: I1014 15:42:47.713218 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:42:47 crc kubenswrapper[4945]: E1014 15:42:47.713560 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:42:56 crc kubenswrapper[4945]: I1014 15:42:56.041199 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-8xv6t"] Oct 14 15:42:56 crc kubenswrapper[4945]: I1014 15:42:56.050507 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-6pgfh"] Oct 14 15:42:56 crc kubenswrapper[4945]: I1014 15:42:56.060775 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-8xv6t"] Oct 14 15:42:56 crc kubenswrapper[4945]: I1014 15:42:56.068750 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-6pgfh"] Oct 14 15:42:56 crc kubenswrapper[4945]: I1014 15:42:56.772336 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8233982e-1252-4dae-a15c-cbdc04652502" path="/var/lib/kubelet/pods/8233982e-1252-4dae-a15c-cbdc04652502/volumes" Oct 14 15:42:56 crc kubenswrapper[4945]: I1014 15:42:56.773375 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e71fad98-9eac-426c-9bc5-5d3776a63a8e" path="/var/lib/kubelet/pods/e71fad98-9eac-426c-9bc5-5d3776a63a8e/volumes" Oct 14 15:42:58 crc kubenswrapper[4945]: I1014 15:42:58.028944 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-8924m"] Oct 14 15:42:58 crc kubenswrapper[4945]: I1014 15:42:58.037010 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-8924m"] Oct 14 15:42:58 crc kubenswrapper[4945]: I1014 15:42:58.048207 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-rw779"] Oct 14 15:42:58 crc kubenswrapper[4945]: I1014 15:42:58.055302 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-rw779"] Oct 14 15:42:58 crc kubenswrapper[4945]: I1014 15:42:58.774520 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60f0f22a-507b-44f9-97d9-bed2a2d6d43e" path="/var/lib/kubelet/pods/60f0f22a-507b-44f9-97d9-bed2a2d6d43e/volumes" Oct 14 15:42:58 crc kubenswrapper[4945]: I1014 15:42:58.775448 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea1a8e32-d4a4-46d5-bedf-5883dbf9761e" path="/var/lib/kubelet/pods/ea1a8e32-d4a4-46d5-bedf-5883dbf9761e/volumes" Oct 14 15:43:02 crc kubenswrapper[4945]: I1014 15:43:02.762769 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:43:02 crc kubenswrapper[4945]: E1014 15:43:02.763354 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:43:11 crc kubenswrapper[4945]: I1014 15:43:11.905066 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8ffg6/must-gather-sld2n"] Oct 14 15:43:11 crc kubenswrapper[4945]: E1014 15:43:11.906061 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" containerName="extract-utilities" Oct 14 15:43:11 crc kubenswrapper[4945]: I1014 15:43:11.906078 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" containerName="extract-utilities" Oct 14 15:43:11 crc kubenswrapper[4945]: E1014 15:43:11.906094 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" containerName="extract-content" Oct 14 15:43:11 crc kubenswrapper[4945]: I1014 15:43:11.906101 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" containerName="extract-content" Oct 14 15:43:11 crc kubenswrapper[4945]: E1014 15:43:11.906118 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" containerName="registry-server" Oct 14 15:43:11 crc kubenswrapper[4945]: I1014 15:43:11.906124 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" containerName="registry-server" Oct 14 15:43:11 crc kubenswrapper[4945]: I1014 15:43:11.906334 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9bc5f5d-9f3d-46ed-a019-a3977adc9da5" containerName="registry-server" Oct 14 15:43:11 crc kubenswrapper[4945]: I1014 15:43:11.907423 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/must-gather-sld2n" Oct 14 15:43:11 crc kubenswrapper[4945]: I1014 15:43:11.914729 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8ffg6"/"openshift-service-ca.crt" Oct 14 15:43:11 crc kubenswrapper[4945]: I1014 15:43:11.914745 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8ffg6"/"kube-root-ca.crt" Oct 14 15:43:11 crc kubenswrapper[4945]: I1014 15:43:11.914973 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8ffg6"/"default-dockercfg-qtpnm" Oct 14 15:43:11 crc kubenswrapper[4945]: I1014 15:43:11.926748 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8ffg6/must-gather-sld2n"] Oct 14 15:43:11 crc kubenswrapper[4945]: I1014 15:43:11.958967 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv9c4\" (UniqueName: \"kubernetes.io/projected/4b474496-81ee-4c0f-b783-6550dd51b361-kube-api-access-cv9c4\") pod \"must-gather-sld2n\" (UID: \"4b474496-81ee-4c0f-b783-6550dd51b361\") " pod="openshift-must-gather-8ffg6/must-gather-sld2n" Oct 14 15:43:11 crc kubenswrapper[4945]: I1014 15:43:11.959036 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4b474496-81ee-4c0f-b783-6550dd51b361-must-gather-output\") pod \"must-gather-sld2n\" (UID: \"4b474496-81ee-4c0f-b783-6550dd51b361\") " pod="openshift-must-gather-8ffg6/must-gather-sld2n" Oct 14 15:43:12 crc kubenswrapper[4945]: I1014 15:43:12.060952 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv9c4\" (UniqueName: \"kubernetes.io/projected/4b474496-81ee-4c0f-b783-6550dd51b361-kube-api-access-cv9c4\") pod \"must-gather-sld2n\" (UID: \"4b474496-81ee-4c0f-b783-6550dd51b361\") " pod="openshift-must-gather-8ffg6/must-gather-sld2n" Oct 14 15:43:12 crc kubenswrapper[4945]: I1014 15:43:12.061011 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4b474496-81ee-4c0f-b783-6550dd51b361-must-gather-output\") pod \"must-gather-sld2n\" (UID: \"4b474496-81ee-4c0f-b783-6550dd51b361\") " pod="openshift-must-gather-8ffg6/must-gather-sld2n" Oct 14 15:43:12 crc kubenswrapper[4945]: I1014 15:43:12.061517 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4b474496-81ee-4c0f-b783-6550dd51b361-must-gather-output\") pod \"must-gather-sld2n\" (UID: \"4b474496-81ee-4c0f-b783-6550dd51b361\") " pod="openshift-must-gather-8ffg6/must-gather-sld2n" Oct 14 15:43:12 crc kubenswrapper[4945]: I1014 15:43:12.087007 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv9c4\" (UniqueName: \"kubernetes.io/projected/4b474496-81ee-4c0f-b783-6550dd51b361-kube-api-access-cv9c4\") pod \"must-gather-sld2n\" (UID: \"4b474496-81ee-4c0f-b783-6550dd51b361\") " pod="openshift-must-gather-8ffg6/must-gather-sld2n" Oct 14 15:43:12 crc kubenswrapper[4945]: I1014 15:43:12.230836 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/must-gather-sld2n" Oct 14 15:43:12 crc kubenswrapper[4945]: I1014 15:43:12.701213 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8ffg6/must-gather-sld2n"] Oct 14 15:43:12 crc kubenswrapper[4945]: I1014 15:43:12.742161 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 15:43:12 crc kubenswrapper[4945]: I1014 15:43:12.934950 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ffg6/must-gather-sld2n" event={"ID":"4b474496-81ee-4c0f-b783-6550dd51b361","Type":"ContainerStarted","Data":"ee5ba5aa1025e42646e04b88fb8ae9d78df4693c089d96d3f54ef86aa7cb7999"} Oct 14 15:43:14 crc kubenswrapper[4945]: I1014 15:43:14.764051 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:43:14 crc kubenswrapper[4945]: E1014 15:43:14.764547 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:43:16 crc kubenswrapper[4945]: I1014 15:43:16.991761 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ffg6/must-gather-sld2n" event={"ID":"4b474496-81ee-4c0f-b783-6550dd51b361","Type":"ContainerStarted","Data":"f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b"} Oct 14 15:43:18 crc kubenswrapper[4945]: I1014 15:43:18.002025 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ffg6/must-gather-sld2n" event={"ID":"4b474496-81ee-4c0f-b783-6550dd51b361","Type":"ContainerStarted","Data":"cdb6459ec5df1973d3a02d957e0b250055fe0d254683d380f362c70d059bade0"} Oct 14 15:43:20 crc kubenswrapper[4945]: I1014 15:43:20.570086 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8ffg6/must-gather-sld2n" podStartSLOduration=5.645740444 podStartE2EDuration="9.570048972s" podCreationTimestamp="2025-10-14 15:43:11 +0000 UTC" firstStartedPulling="2025-10-14 15:43:12.741936121 +0000 UTC m=+1722.725984489" lastFinishedPulling="2025-10-14 15:43:16.666244639 +0000 UTC m=+1726.650293017" observedRunningTime="2025-10-14 15:43:18.019860869 +0000 UTC m=+1728.003909237" watchObservedRunningTime="2025-10-14 15:43:20.570048972 +0000 UTC m=+1730.554097340" Oct 14 15:43:20 crc kubenswrapper[4945]: I1014 15:43:20.581805 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8ffg6/crc-debug-hx6m5"] Oct 14 15:43:20 crc kubenswrapper[4945]: I1014 15:43:20.583424 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" Oct 14 15:43:20 crc kubenswrapper[4945]: I1014 15:43:20.593182 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qhxm\" (UniqueName: \"kubernetes.io/projected/b3071f9d-9717-42bd-8084-f83717f6cb4e-kube-api-access-2qhxm\") pod \"crc-debug-hx6m5\" (UID: \"b3071f9d-9717-42bd-8084-f83717f6cb4e\") " pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" Oct 14 15:43:20 crc kubenswrapper[4945]: I1014 15:43:20.593267 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b3071f9d-9717-42bd-8084-f83717f6cb4e-host\") pod \"crc-debug-hx6m5\" (UID: \"b3071f9d-9717-42bd-8084-f83717f6cb4e\") " pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" Oct 14 15:43:20 crc kubenswrapper[4945]: I1014 15:43:20.694560 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qhxm\" (UniqueName: \"kubernetes.io/projected/b3071f9d-9717-42bd-8084-f83717f6cb4e-kube-api-access-2qhxm\") pod \"crc-debug-hx6m5\" (UID: \"b3071f9d-9717-42bd-8084-f83717f6cb4e\") " pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" Oct 14 15:43:20 crc kubenswrapper[4945]: I1014 15:43:20.694654 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b3071f9d-9717-42bd-8084-f83717f6cb4e-host\") pod \"crc-debug-hx6m5\" (UID: \"b3071f9d-9717-42bd-8084-f83717f6cb4e\") " pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" Oct 14 15:43:20 crc kubenswrapper[4945]: I1014 15:43:20.694834 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b3071f9d-9717-42bd-8084-f83717f6cb4e-host\") pod \"crc-debug-hx6m5\" (UID: \"b3071f9d-9717-42bd-8084-f83717f6cb4e\") " pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" Oct 14 15:43:20 crc kubenswrapper[4945]: I1014 15:43:20.715159 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qhxm\" (UniqueName: \"kubernetes.io/projected/b3071f9d-9717-42bd-8084-f83717f6cb4e-kube-api-access-2qhxm\") pod \"crc-debug-hx6m5\" (UID: \"b3071f9d-9717-42bd-8084-f83717f6cb4e\") " pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" Oct 14 15:43:20 crc kubenswrapper[4945]: I1014 15:43:20.906570 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" Oct 14 15:43:20 crc kubenswrapper[4945]: W1014 15:43:20.959541 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3071f9d_9717_42bd_8084_f83717f6cb4e.slice/crio-db591e92b93a169eca5cc6565426bf6cf1a0710b9a3ad1e00e4c9cbd1aecd459 WatchSource:0}: Error finding container db591e92b93a169eca5cc6565426bf6cf1a0710b9a3ad1e00e4c9cbd1aecd459: Status 404 returned error can't find the container with id db591e92b93a169eca5cc6565426bf6cf1a0710b9a3ad1e00e4c9cbd1aecd459 Oct 14 15:43:21 crc kubenswrapper[4945]: I1014 15:43:21.027708 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" event={"ID":"b3071f9d-9717-42bd-8084-f83717f6cb4e","Type":"ContainerStarted","Data":"db591e92b93a169eca5cc6565426bf6cf1a0710b9a3ad1e00e4c9cbd1aecd459"} Oct 14 15:43:26 crc kubenswrapper[4945]: I1014 15:43:26.763439 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:43:26 crc kubenswrapper[4945]: E1014 15:43:26.764208 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:43:33 crc kubenswrapper[4945]: I1014 15:43:33.140697 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" event={"ID":"b3071f9d-9717-42bd-8084-f83717f6cb4e","Type":"ContainerStarted","Data":"1e2872b7355c57f813dfa8bec061238ec7b3bce856ead60ed2ebe2fc124e69cc"} Oct 14 15:43:33 crc kubenswrapper[4945]: I1014 15:43:33.156816 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" podStartSLOduration=1.336250207 podStartE2EDuration="13.156793938s" podCreationTimestamp="2025-10-14 15:43:20 +0000 UTC" firstStartedPulling="2025-10-14 15:43:20.96160696 +0000 UTC m=+1730.945655328" lastFinishedPulling="2025-10-14 15:43:32.782150691 +0000 UTC m=+1742.766199059" observedRunningTime="2025-10-14 15:43:33.153794722 +0000 UTC m=+1743.137843110" watchObservedRunningTime="2025-10-14 15:43:33.156793938 +0000 UTC m=+1743.140842296" Oct 14 15:43:35 crc kubenswrapper[4945]: I1014 15:43:35.038029 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-db-create-8hrxc"] Oct 14 15:43:35 crc kubenswrapper[4945]: I1014 15:43:35.046631 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-db-create-8hrxc"] Oct 14 15:43:36 crc kubenswrapper[4945]: I1014 15:43:36.773649 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92ed91ce-91c5-46ff-8e6c-959b9fa96552" path="/var/lib/kubelet/pods/92ed91ce-91c5-46ff-8e6c-959b9fa96552/volumes" Oct 14 15:43:41 crc kubenswrapper[4945]: I1014 15:43:41.762906 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:43:41 crc kubenswrapper[4945]: E1014 15:43:41.763733 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:43:44 crc kubenswrapper[4945]: I1014 15:43:44.755665 4945 scope.go:117] "RemoveContainer" containerID="de1877f5aac5307e686a80303477a53ed7ad88939c78a0508ec5ac981fe4518b" Oct 14 15:43:45 crc kubenswrapper[4945]: I1014 15:43:45.030549 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-c7c3-account-create-pz2vd"] Oct 14 15:43:45 crc kubenswrapper[4945]: I1014 15:43:45.037782 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-c7c3-account-create-pz2vd"] Oct 14 15:43:45 crc kubenswrapper[4945]: I1014 15:43:45.442967 4945 scope.go:117] "RemoveContainer" containerID="778e15952d386fbbecd7166dac417a95d35853e9b3660d8743a32a5cc8604e67" Oct 14 15:43:45 crc kubenswrapper[4945]: I1014 15:43:45.513255 4945 scope.go:117] "RemoveContainer" containerID="86e788c8e24d13173990293bf8dd21fd10234f901f245aba9028807f7451108d" Oct 14 15:43:45 crc kubenswrapper[4945]: I1014 15:43:45.559664 4945 scope.go:117] "RemoveContainer" containerID="53fc5f52d99e139f5e2baab20bc3a4520e33ed8f4a12f26e927ddf2edcdae561" Oct 14 15:43:45 crc kubenswrapper[4945]: I1014 15:43:45.615015 4945 scope.go:117] "RemoveContainer" containerID="5f0cf2932656c6e5688adbf9fac5a3175e98d84306a5935b0046850b86a9e240" Oct 14 15:43:46 crc kubenswrapper[4945]: I1014 15:43:46.775064 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d9430f4-426a-40f2-8c6d-c9ce794d48a3" path="/var/lib/kubelet/pods/6d9430f4-426a-40f2-8c6d-c9ce794d48a3/volumes" Oct 14 15:43:50 crc kubenswrapper[4945]: I1014 15:43:50.027619 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-zn84r"] Oct 14 15:43:50 crc kubenswrapper[4945]: I1014 15:43:50.037898 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-k49f8"] Oct 14 15:43:50 crc kubenswrapper[4945]: I1014 15:43:50.045251 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-xqd5f"] Oct 14 15:43:50 crc kubenswrapper[4945]: I1014 15:43:50.052024 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-zn84r"] Oct 14 15:43:50 crc kubenswrapper[4945]: I1014 15:43:50.058681 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-xqd5f"] Oct 14 15:43:50 crc kubenswrapper[4945]: I1014 15:43:50.065359 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-k49f8"] Oct 14 15:43:50 crc kubenswrapper[4945]: I1014 15:43:50.773282 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d684cba-7b02-453e-8fc0-f2607daa9951" path="/var/lib/kubelet/pods/5d684cba-7b02-453e-8fc0-f2607daa9951/volumes" Oct 14 15:43:50 crc kubenswrapper[4945]: I1014 15:43:50.773824 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="874d6996-5c8a-4be4-b3e0-ed7a102cf8ba" path="/var/lib/kubelet/pods/874d6996-5c8a-4be4-b3e0-ed7a102cf8ba/volumes" Oct 14 15:43:50 crc kubenswrapper[4945]: I1014 15:43:50.774341 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae872229-e7d9-4fcd-b22d-650f80e8f18a" path="/var/lib/kubelet/pods/ae872229-e7d9-4fcd-b22d-650f80e8f18a/volumes" Oct 14 15:43:53 crc kubenswrapper[4945]: I1014 15:43:53.763394 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:43:53 crc kubenswrapper[4945]: E1014 15:43:53.764027 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:44:01 crc kubenswrapper[4945]: I1014 15:44:01.055476 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-4571-account-create-r2t9d"] Oct 14 15:44:01 crc kubenswrapper[4945]: I1014 15:44:01.063484 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ce31-account-create-425x4"] Oct 14 15:44:01 crc kubenswrapper[4945]: I1014 15:44:01.071432 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cdc7-account-create-qqwb7"] Oct 14 15:44:01 crc kubenswrapper[4945]: I1014 15:44:01.082471 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cdc7-account-create-qqwb7"] Oct 14 15:44:01 crc kubenswrapper[4945]: I1014 15:44:01.089823 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-4571-account-create-r2t9d"] Oct 14 15:44:01 crc kubenswrapper[4945]: I1014 15:44:01.098472 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ce31-account-create-425x4"] Oct 14 15:44:02 crc kubenswrapper[4945]: I1014 15:44:02.773639 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82992e17-467c-46cf-8348-4248a6844b71" path="/var/lib/kubelet/pods/82992e17-467c-46cf-8348-4248a6844b71/volumes" Oct 14 15:44:02 crc kubenswrapper[4945]: I1014 15:44:02.774438 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dbf2952-a679-4420-910d-0ec3b3d6e1c1" path="/var/lib/kubelet/pods/9dbf2952-a679-4420-910d-0ec3b3d6e1c1/volumes" Oct 14 15:44:02 crc kubenswrapper[4945]: I1014 15:44:02.774954 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0c8cdd9-c134-4013-b110-388be1ead4db" path="/var/lib/kubelet/pods/d0c8cdd9-c134-4013-b110-388be1ead4db/volumes" Oct 14 15:44:08 crc kubenswrapper[4945]: I1014 15:44:08.763322 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:44:08 crc kubenswrapper[4945]: E1014 15:44:08.764028 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:44:12 crc kubenswrapper[4945]: I1014 15:44:12.512437 4945 generic.go:334] "Generic (PLEG): container finished" podID="b3071f9d-9717-42bd-8084-f83717f6cb4e" containerID="1e2872b7355c57f813dfa8bec061238ec7b3bce856ead60ed2ebe2fc124e69cc" exitCode=0 Oct 14 15:44:12 crc kubenswrapper[4945]: I1014 15:44:12.512536 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" event={"ID":"b3071f9d-9717-42bd-8084-f83717f6cb4e","Type":"ContainerDied","Data":"1e2872b7355c57f813dfa8bec061238ec7b3bce856ead60ed2ebe2fc124e69cc"} Oct 14 15:44:13 crc kubenswrapper[4945]: I1014 15:44:13.625711 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" Oct 14 15:44:13 crc kubenswrapper[4945]: I1014 15:44:13.667535 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8ffg6/crc-debug-hx6m5"] Oct 14 15:44:13 crc kubenswrapper[4945]: I1014 15:44:13.677102 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8ffg6/crc-debug-hx6m5"] Oct 14 15:44:13 crc kubenswrapper[4945]: I1014 15:44:13.761958 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qhxm\" (UniqueName: \"kubernetes.io/projected/b3071f9d-9717-42bd-8084-f83717f6cb4e-kube-api-access-2qhxm\") pod \"b3071f9d-9717-42bd-8084-f83717f6cb4e\" (UID: \"b3071f9d-9717-42bd-8084-f83717f6cb4e\") " Oct 14 15:44:13 crc kubenswrapper[4945]: I1014 15:44:13.762271 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b3071f9d-9717-42bd-8084-f83717f6cb4e-host\") pod \"b3071f9d-9717-42bd-8084-f83717f6cb4e\" (UID: \"b3071f9d-9717-42bd-8084-f83717f6cb4e\") " Oct 14 15:44:13 crc kubenswrapper[4945]: I1014 15:44:13.762449 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3071f9d-9717-42bd-8084-f83717f6cb4e-host" (OuterVolumeSpecName: "host") pod "b3071f9d-9717-42bd-8084-f83717f6cb4e" (UID: "b3071f9d-9717-42bd-8084-f83717f6cb4e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:44:13 crc kubenswrapper[4945]: I1014 15:44:13.763151 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b3071f9d-9717-42bd-8084-f83717f6cb4e-host\") on node \"crc\" DevicePath \"\"" Oct 14 15:44:13 crc kubenswrapper[4945]: I1014 15:44:13.782185 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3071f9d-9717-42bd-8084-f83717f6cb4e-kube-api-access-2qhxm" (OuterVolumeSpecName: "kube-api-access-2qhxm") pod "b3071f9d-9717-42bd-8084-f83717f6cb4e" (UID: "b3071f9d-9717-42bd-8084-f83717f6cb4e"). InnerVolumeSpecName "kube-api-access-2qhxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:44:13 crc kubenswrapper[4945]: I1014 15:44:13.865065 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qhxm\" (UniqueName: \"kubernetes.io/projected/b3071f9d-9717-42bd-8084-f83717f6cb4e-kube-api-access-2qhxm\") on node \"crc\" DevicePath \"\"" Oct 14 15:44:14 crc kubenswrapper[4945]: I1014 15:44:14.541144 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db591e92b93a169eca5cc6565426bf6cf1a0710b9a3ad1e00e4c9cbd1aecd459" Oct 14 15:44:14 crc kubenswrapper[4945]: I1014 15:44:14.541183 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/crc-debug-hx6m5" Oct 14 15:44:14 crc kubenswrapper[4945]: I1014 15:44:14.788045 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3071f9d-9717-42bd-8084-f83717f6cb4e" path="/var/lib/kubelet/pods/b3071f9d-9717-42bd-8084-f83717f6cb4e/volumes" Oct 14 15:44:14 crc kubenswrapper[4945]: I1014 15:44:14.828885 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8ffg6/crc-debug-g9dsg"] Oct 14 15:44:14 crc kubenswrapper[4945]: E1014 15:44:14.829254 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3071f9d-9717-42bd-8084-f83717f6cb4e" containerName="container-00" Oct 14 15:44:14 crc kubenswrapper[4945]: I1014 15:44:14.829271 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3071f9d-9717-42bd-8084-f83717f6cb4e" containerName="container-00" Oct 14 15:44:14 crc kubenswrapper[4945]: I1014 15:44:14.829456 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3071f9d-9717-42bd-8084-f83717f6cb4e" containerName="container-00" Oct 14 15:44:14 crc kubenswrapper[4945]: I1014 15:44:14.830030 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/crc-debug-g9dsg" Oct 14 15:44:14 crc kubenswrapper[4945]: I1014 15:44:14.891409 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d47jt\" (UniqueName: \"kubernetes.io/projected/726d9555-abcf-443b-8260-c5c2f3f5331a-kube-api-access-d47jt\") pod \"crc-debug-g9dsg\" (UID: \"726d9555-abcf-443b-8260-c5c2f3f5331a\") " pod="openshift-must-gather-8ffg6/crc-debug-g9dsg" Oct 14 15:44:14 crc kubenswrapper[4945]: I1014 15:44:14.891461 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/726d9555-abcf-443b-8260-c5c2f3f5331a-host\") pod \"crc-debug-g9dsg\" (UID: \"726d9555-abcf-443b-8260-c5c2f3f5331a\") " pod="openshift-must-gather-8ffg6/crc-debug-g9dsg" Oct 14 15:44:14 crc kubenswrapper[4945]: I1014 15:44:14.993711 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d47jt\" (UniqueName: \"kubernetes.io/projected/726d9555-abcf-443b-8260-c5c2f3f5331a-kube-api-access-d47jt\") pod \"crc-debug-g9dsg\" (UID: \"726d9555-abcf-443b-8260-c5c2f3f5331a\") " pod="openshift-must-gather-8ffg6/crc-debug-g9dsg" Oct 14 15:44:14 crc kubenswrapper[4945]: I1014 15:44:14.993770 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/726d9555-abcf-443b-8260-c5c2f3f5331a-host\") pod \"crc-debug-g9dsg\" (UID: \"726d9555-abcf-443b-8260-c5c2f3f5331a\") " pod="openshift-must-gather-8ffg6/crc-debug-g9dsg" Oct 14 15:44:14 crc kubenswrapper[4945]: I1014 15:44:14.993981 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/726d9555-abcf-443b-8260-c5c2f3f5331a-host\") pod \"crc-debug-g9dsg\" (UID: \"726d9555-abcf-443b-8260-c5c2f3f5331a\") " pod="openshift-must-gather-8ffg6/crc-debug-g9dsg" Oct 14 15:44:15 crc kubenswrapper[4945]: I1014 15:44:15.013198 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d47jt\" (UniqueName: \"kubernetes.io/projected/726d9555-abcf-443b-8260-c5c2f3f5331a-kube-api-access-d47jt\") pod \"crc-debug-g9dsg\" (UID: \"726d9555-abcf-443b-8260-c5c2f3f5331a\") " pod="openshift-must-gather-8ffg6/crc-debug-g9dsg" Oct 14 15:44:15 crc kubenswrapper[4945]: I1014 15:44:15.154182 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/crc-debug-g9dsg" Oct 14 15:44:15 crc kubenswrapper[4945]: I1014 15:44:15.553105 4945 generic.go:334] "Generic (PLEG): container finished" podID="726d9555-abcf-443b-8260-c5c2f3f5331a" containerID="d4189e827eec93ab1328aa0bfd9ec1680dc82fa6da5ae34600784ea21196966c" exitCode=0 Oct 14 15:44:15 crc kubenswrapper[4945]: I1014 15:44:15.553172 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ffg6/crc-debug-g9dsg" event={"ID":"726d9555-abcf-443b-8260-c5c2f3f5331a","Type":"ContainerDied","Data":"d4189e827eec93ab1328aa0bfd9ec1680dc82fa6da5ae34600784ea21196966c"} Oct 14 15:44:15 crc kubenswrapper[4945]: I1014 15:44:15.553215 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ffg6/crc-debug-g9dsg" event={"ID":"726d9555-abcf-443b-8260-c5c2f3f5331a","Type":"ContainerStarted","Data":"eb40dcbafdca0d7af5f1199841b03dc49bc7e5aa42749e28f7d8a8bba4764e19"} Oct 14 15:44:16 crc kubenswrapper[4945]: I1014 15:44:16.062089 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8ffg6/crc-debug-g9dsg"] Oct 14 15:44:16 crc kubenswrapper[4945]: I1014 15:44:16.069646 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8ffg6/crc-debug-g9dsg"] Oct 14 15:44:16 crc kubenswrapper[4945]: I1014 15:44:16.661302 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/crc-debug-g9dsg" Oct 14 15:44:16 crc kubenswrapper[4945]: I1014 15:44:16.724588 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/726d9555-abcf-443b-8260-c5c2f3f5331a-host\") pod \"726d9555-abcf-443b-8260-c5c2f3f5331a\" (UID: \"726d9555-abcf-443b-8260-c5c2f3f5331a\") " Oct 14 15:44:16 crc kubenswrapper[4945]: I1014 15:44:16.724646 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d47jt\" (UniqueName: \"kubernetes.io/projected/726d9555-abcf-443b-8260-c5c2f3f5331a-kube-api-access-d47jt\") pod \"726d9555-abcf-443b-8260-c5c2f3f5331a\" (UID: \"726d9555-abcf-443b-8260-c5c2f3f5331a\") " Oct 14 15:44:16 crc kubenswrapper[4945]: I1014 15:44:16.724742 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/726d9555-abcf-443b-8260-c5c2f3f5331a-host" (OuterVolumeSpecName: "host") pod "726d9555-abcf-443b-8260-c5c2f3f5331a" (UID: "726d9555-abcf-443b-8260-c5c2f3f5331a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:44:16 crc kubenswrapper[4945]: I1014 15:44:16.725183 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/726d9555-abcf-443b-8260-c5c2f3f5331a-host\") on node \"crc\" DevicePath \"\"" Oct 14 15:44:16 crc kubenswrapper[4945]: I1014 15:44:16.737112 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/726d9555-abcf-443b-8260-c5c2f3f5331a-kube-api-access-d47jt" (OuterVolumeSpecName: "kube-api-access-d47jt") pod "726d9555-abcf-443b-8260-c5c2f3f5331a" (UID: "726d9555-abcf-443b-8260-c5c2f3f5331a"). InnerVolumeSpecName "kube-api-access-d47jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:44:16 crc kubenswrapper[4945]: I1014 15:44:16.776614 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="726d9555-abcf-443b-8260-c5c2f3f5331a" path="/var/lib/kubelet/pods/726d9555-abcf-443b-8260-c5c2f3f5331a/volumes" Oct 14 15:44:16 crc kubenswrapper[4945]: I1014 15:44:16.827444 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d47jt\" (UniqueName: \"kubernetes.io/projected/726d9555-abcf-443b-8260-c5c2f3f5331a-kube-api-access-d47jt\") on node \"crc\" DevicePath \"\"" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.281599 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8ffg6/crc-debug-ln5qg"] Oct 14 15:44:17 crc kubenswrapper[4945]: E1014 15:44:17.282036 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="726d9555-abcf-443b-8260-c5c2f3f5331a" containerName="container-00" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.282049 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="726d9555-abcf-443b-8260-c5c2f3f5331a" containerName="container-00" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.282244 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="726d9555-abcf-443b-8260-c5c2f3f5331a" containerName="container-00" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.282842 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/crc-debug-ln5qg" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.336686 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8stfc\" (UniqueName: \"kubernetes.io/projected/195282c5-c050-4a90-b4cb-97f612fa3286-kube-api-access-8stfc\") pod \"crc-debug-ln5qg\" (UID: \"195282c5-c050-4a90-b4cb-97f612fa3286\") " pod="openshift-must-gather-8ffg6/crc-debug-ln5qg" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.336857 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/195282c5-c050-4a90-b4cb-97f612fa3286-host\") pod \"crc-debug-ln5qg\" (UID: \"195282c5-c050-4a90-b4cb-97f612fa3286\") " pod="openshift-must-gather-8ffg6/crc-debug-ln5qg" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.438688 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8stfc\" (UniqueName: \"kubernetes.io/projected/195282c5-c050-4a90-b4cb-97f612fa3286-kube-api-access-8stfc\") pod \"crc-debug-ln5qg\" (UID: \"195282c5-c050-4a90-b4cb-97f612fa3286\") " pod="openshift-must-gather-8ffg6/crc-debug-ln5qg" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.439168 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/195282c5-c050-4a90-b4cb-97f612fa3286-host\") pod \"crc-debug-ln5qg\" (UID: \"195282c5-c050-4a90-b4cb-97f612fa3286\") " pod="openshift-must-gather-8ffg6/crc-debug-ln5qg" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.439353 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/195282c5-c050-4a90-b4cb-97f612fa3286-host\") pod \"crc-debug-ln5qg\" (UID: \"195282c5-c050-4a90-b4cb-97f612fa3286\") " pod="openshift-must-gather-8ffg6/crc-debug-ln5qg" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.458775 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8stfc\" (UniqueName: \"kubernetes.io/projected/195282c5-c050-4a90-b4cb-97f612fa3286-kube-api-access-8stfc\") pod \"crc-debug-ln5qg\" (UID: \"195282c5-c050-4a90-b4cb-97f612fa3286\") " pod="openshift-must-gather-8ffg6/crc-debug-ln5qg" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.570463 4945 scope.go:117] "RemoveContainer" containerID="d4189e827eec93ab1328aa0bfd9ec1680dc82fa6da5ae34600784ea21196966c" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.570507 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/crc-debug-g9dsg" Oct 14 15:44:17 crc kubenswrapper[4945]: I1014 15:44:17.600597 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/crc-debug-ln5qg" Oct 14 15:44:17 crc kubenswrapper[4945]: W1014 15:44:17.629701 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod195282c5_c050_4a90_b4cb_97f612fa3286.slice/crio-d82ccef9947af18772c23735c24b20c03c2ee659c48733c7cea7fefd571f7084 WatchSource:0}: Error finding container d82ccef9947af18772c23735c24b20c03c2ee659c48733c7cea7fefd571f7084: Status 404 returned error can't find the container with id d82ccef9947af18772c23735c24b20c03c2ee659c48733c7cea7fefd571f7084 Oct 14 15:44:18 crc kubenswrapper[4945]: I1014 15:44:18.551506 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7f6f97bcf4-qbb2j_9615a643-2f07-4637-b8d0-3acb4f24d0d8/barbican-api-log/0.log" Oct 14 15:44:18 crc kubenswrapper[4945]: I1014 15:44:18.590167 4945 generic.go:334] "Generic (PLEG): container finished" podID="195282c5-c050-4a90-b4cb-97f612fa3286" containerID="44e7ee7e570b62a6097705ab4abaadd6c176e9546a222b4a2610ca9faa443af8" exitCode=0 Oct 14 15:44:18 crc kubenswrapper[4945]: I1014 15:44:18.590231 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ffg6/crc-debug-ln5qg" event={"ID":"195282c5-c050-4a90-b4cb-97f612fa3286","Type":"ContainerDied","Data":"44e7ee7e570b62a6097705ab4abaadd6c176e9546a222b4a2610ca9faa443af8"} Oct 14 15:44:18 crc kubenswrapper[4945]: I1014 15:44:18.590286 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ffg6/crc-debug-ln5qg" event={"ID":"195282c5-c050-4a90-b4cb-97f612fa3286","Type":"ContainerStarted","Data":"d82ccef9947af18772c23735c24b20c03c2ee659c48733c7cea7fefd571f7084"} Oct 14 15:44:18 crc kubenswrapper[4945]: I1014 15:44:18.603492 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7f6f97bcf4-qbb2j_9615a643-2f07-4637-b8d0-3acb4f24d0d8/barbican-api/0.log" Oct 14 15:44:18 crc kubenswrapper[4945]: I1014 15:44:18.634859 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8ffg6/crc-debug-ln5qg"] Oct 14 15:44:18 crc kubenswrapper[4945]: I1014 15:44:18.642793 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8ffg6/crc-debug-ln5qg"] Oct 14 15:44:18 crc kubenswrapper[4945]: I1014 15:44:18.792715 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-dfb99d5c8-zqfxx_e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67/barbican-keystone-listener/0.log" Oct 14 15:44:18 crc kubenswrapper[4945]: I1014 15:44:18.945230 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-dfb99d5c8-zqfxx_e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67/barbican-keystone-listener-log/0.log" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.015740 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5749bb9b55-8z9wq_fcc68ba8-49a8-456d-b808-2d63dc0d6377/barbican-worker/0.log" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.050015 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5749bb9b55-8z9wq_fcc68ba8-49a8-456d-b808-2d63dc0d6377/barbican-worker-log/0.log" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.226306 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_519a62f5-6d77-4b41-9d34-52651d4e9571/ceilometer-notification-agent/0.log" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.234383 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_519a62f5-6d77-4b41-9d34-52651d4e9571/ceilometer-central-agent/0.log" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.324691 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_519a62f5-6d77-4b41-9d34-52651d4e9571/proxy-httpd/0.log" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.422577 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_519a62f5-6d77-4b41-9d34-52651d4e9571/sg-core/0.log" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.453152 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b210a0b1-468b-4e67-8618-02cb27c78033/cinder-api/0.log" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.561786 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b210a0b1-468b-4e67-8618-02cb27c78033/cinder-api-log/0.log" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.662668 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b15e8808-64a4-4e83-8a8a-1ff4ecc68711/cinder-scheduler/0.log" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.685674 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b15e8808-64a4-4e83-8a8a-1ff4ecc68711/probe/0.log" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.728131 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/crc-debug-ln5qg" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.790330 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8stfc\" (UniqueName: \"kubernetes.io/projected/195282c5-c050-4a90-b4cb-97f612fa3286-kube-api-access-8stfc\") pod \"195282c5-c050-4a90-b4cb-97f612fa3286\" (UID: \"195282c5-c050-4a90-b4cb-97f612fa3286\") " Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.790606 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/195282c5-c050-4a90-b4cb-97f612fa3286-host\") pod \"195282c5-c050-4a90-b4cb-97f612fa3286\" (UID: \"195282c5-c050-4a90-b4cb-97f612fa3286\") " Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.790689 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/195282c5-c050-4a90-b4cb-97f612fa3286-host" (OuterVolumeSpecName: "host") pod "195282c5-c050-4a90-b4cb-97f612fa3286" (UID: "195282c5-c050-4a90-b4cb-97f612fa3286"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.790997 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/195282c5-c050-4a90-b4cb-97f612fa3286-host\") on node \"crc\" DevicePath \"\"" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.796119 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/195282c5-c050-4a90-b4cb-97f612fa3286-kube-api-access-8stfc" (OuterVolumeSpecName: "kube-api-access-8stfc") pod "195282c5-c050-4a90-b4cb-97f612fa3286" (UID: "195282c5-c050-4a90-b4cb-97f612fa3286"). InnerVolumeSpecName "kube-api-access-8stfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.868110 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5c7b6c5df9-hhtr4_9b5e61fe-2a22-4128-90c3-569e2e83a972/init/0.log" Oct 14 15:44:19 crc kubenswrapper[4945]: I1014 15:44:19.893332 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8stfc\" (UniqueName: \"kubernetes.io/projected/195282c5-c050-4a90-b4cb-97f612fa3286-kube-api-access-8stfc\") on node \"crc\" DevicePath \"\"" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.009812 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5c7b6c5df9-hhtr4_9b5e61fe-2a22-4128-90c3-569e2e83a972/init/0.log" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.074044 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5c7b6c5df9-hhtr4_9b5e61fe-2a22-4128-90c3-569e2e83a972/dnsmasq-dns/0.log" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.105031 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_57cbaab0-b5ce-4581-af58-87e3e3ec679c/glance-httpd/0.log" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.213946 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_57cbaab0-b5ce-4581-af58-87e3e3ec679c/glance-log/0.log" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.257503 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_314804da-3d6d-44cd-a9cf-0f014d690fe9/glance-httpd/0.log" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.298860 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_314804da-3d6d-44cd-a9cf-0f014d690fe9/glance-log/0.log" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.480406 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-bf87b4676-qpt54_5f9f45b7-0e91-4ecf-b38f-49bb07f40868/init/0.log" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.609564 4945 scope.go:117] "RemoveContainer" containerID="44e7ee7e570b62a6097705ab4abaadd6c176e9546a222b4a2610ca9faa443af8" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.609733 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/crc-debug-ln5qg" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.653051 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-bf87b4676-qpt54_5f9f45b7-0e91-4ecf-b38f-49bb07f40868/ironic-api-log/0.log" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.655729 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-bf87b4676-qpt54_5f9f45b7-0e91-4ecf-b38f-49bb07f40868/init/0.log" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.741454 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-bf87b4676-qpt54_5f9f45b7-0e91-4ecf-b38f-49bb07f40868/ironic-api/0.log" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.774481 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="195282c5-c050-4a90-b4cb-97f612fa3286" path="/var/lib/kubelet/pods/195282c5-c050-4a90-b4cb-97f612fa3286/volumes" Oct 14 15:44:20 crc kubenswrapper[4945]: I1014 15:44:20.899712 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/init/0.log" Oct 14 15:44:21 crc kubenswrapper[4945]: I1014 15:44:21.039383 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/init/0.log" Oct 14 15:44:21 crc kubenswrapper[4945]: I1014 15:44:21.060108 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/ironic-python-agent-init/0.log" Oct 14 15:44:21 crc kubenswrapper[4945]: I1014 15:44:21.143488 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/ironic-python-agent-init/0.log" Oct 14 15:44:21 crc kubenswrapper[4945]: I1014 15:44:21.396947 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/init/0.log" Oct 14 15:44:21 crc kubenswrapper[4945]: I1014 15:44:21.436088 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/ironic-python-agent-init/0.log" Oct 14 15:44:21 crc kubenswrapper[4945]: I1014 15:44:21.809351 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/ironic-python-agent-init/0.log" Oct 14 15:44:21 crc kubenswrapper[4945]: I1014 15:44:21.829556 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/init/0.log" Oct 14 15:44:21 crc kubenswrapper[4945]: I1014 15:44:21.874147 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/pxe-init/0.log" Oct 14 15:44:22 crc kubenswrapper[4945]: I1014 15:44:22.096228 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/httpboot/0.log" Oct 14 15:44:22 crc kubenswrapper[4945]: I1014 15:44:22.166466 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/ironic-conductor/0.log" Oct 14 15:44:22 crc kubenswrapper[4945]: I1014 15:44:22.297260 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/ramdisk-logs/0.log" Oct 14 15:44:22 crc kubenswrapper[4945]: I1014 15:44:22.472541 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-gcqrl_367d71fe-5d2e-470a-958d-4a97ad15545c/init/0.log" Oct 14 15:44:22 crc kubenswrapper[4945]: I1014 15:44:22.738382 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/pxe-init/0.log" Oct 14 15:44:22 crc kubenswrapper[4945]: I1014 15:44:22.812974 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-gcqrl_367d71fe-5d2e-470a-958d-4a97ad15545c/init/0.log" Oct 14 15:44:22 crc kubenswrapper[4945]: I1014 15:44:22.953366 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-gcqrl_367d71fe-5d2e-470a-958d-4a97ad15545c/ironic-db-sync/0.log" Oct 14 15:44:22 crc kubenswrapper[4945]: I1014 15:44:22.978080 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/pxe-init/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.121432 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/ironic-python-agent-init/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.138063 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/pxe-init/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.240758 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/ironic-python-agent-init/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.248665 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/inspector-pxe-init/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.313201 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/inspector-pxe-init/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.551073 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/ironic-inspector/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.555671 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/ironic-python-agent-init/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.564265 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/inspector-pxe-init/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.584184 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/inspector-httpboot/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.600033 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/ironic-inspector-httpd/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.762700 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:44:23 crc kubenswrapper[4945]: E1014 15:44:23.763000 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.774223 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/ramdisk-logs/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.789454 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-db-sync-t252w_048f7a68-56de-43d2-8ffb-33acacc4e2ec/ironic-inspector-db-sync/0.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.840179 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-neutron-agent-648bd876cc-srkfr_11506640-cea7-421c-878c-c4b449df880f/ironic-neutron-agent/3.log" Oct 14 15:44:23 crc kubenswrapper[4945]: I1014 15:44:23.961773 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-neutron-agent-648bd876cc-srkfr_11506640-cea7-421c-878c-c4b449df880f/ironic-neutron-agent/2.log" Oct 14 15:44:24 crc kubenswrapper[4945]: I1014 15:44:24.057325 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-h5wlv"] Oct 14 15:44:24 crc kubenswrapper[4945]: I1014 15:44:24.067080 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-h5wlv"] Oct 14 15:44:24 crc kubenswrapper[4945]: I1014 15:44:24.095164 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_41add807-ab1f-4d5f-999f-51626e4bb839/kube-state-metrics/0.log" Oct 14 15:44:24 crc kubenswrapper[4945]: I1014 15:44:24.176765 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5cdf875b4d-nbblw_fab41439-cbf7-47cd-a439-70519713d3db/keystone-api/0.log" Oct 14 15:44:24 crc kubenswrapper[4945]: I1014 15:44:24.455740 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7cb8f5594f-b9dt5_904476f1-4b6d-432f-9db6-a0fd60f6ac3a/neutron-httpd/0.log" Oct 14 15:44:24 crc kubenswrapper[4945]: I1014 15:44:24.481050 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7cb8f5594f-b9dt5_904476f1-4b6d-432f-9db6-a0fd60f6ac3a/neutron-api/0.log" Oct 14 15:44:24 crc kubenswrapper[4945]: I1014 15:44:24.777506 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a821c3a6-1324-4c0e-bcf4-1700cb3cfd69/nova-api-api/0.log" Oct 14 15:44:24 crc kubenswrapper[4945]: I1014 15:44:24.781925 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82939cb9-d9fc-4620-9a17-8990bd27bba8" path="/var/lib/kubelet/pods/82939cb9-d9fc-4620-9a17-8990bd27bba8/volumes" Oct 14 15:44:24 crc kubenswrapper[4945]: I1014 15:44:24.785345 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a821c3a6-1324-4c0e-bcf4-1700cb3cfd69/nova-api-log/0.log" Oct 14 15:44:24 crc kubenswrapper[4945]: I1014 15:44:24.864333 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-cell-mapping-b7scm_63c34001-2d18-4f96-a9e5-af05eab7f11b/nova-manage/0.log" Oct 14 15:44:25 crc kubenswrapper[4945]: I1014 15:44:25.205561 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f4bf8e7b-6b0a-48da-831c-a9c307af75b9/nova-cell0-conductor-conductor/0.log" Oct 14 15:44:25 crc kubenswrapper[4945]: I1014 15:44:25.388466 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-cell-mapping-nvcm2_c2518d4a-a272-4e4e-8c9f-8884d8e4f612/nova-manage/0.log" Oct 14 15:44:25 crc kubenswrapper[4945]: I1014 15:44:25.580064 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_aaedddf9-b122-48e8-9dfc-e3525faaf9d3/nova-cell1-conductor-conductor/0.log" Oct 14 15:44:25 crc kubenswrapper[4945]: I1014 15:44:25.681276 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-db-sync-fwjnr_10a43a64-535b-4cfe-be95-3f6f06d8eb98/nova-cell1-conductor-db-sync/0.log" Oct 14 15:44:25 crc kubenswrapper[4945]: I1014 15:44:25.891616 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9af8e27a-69da-4094-925f-0d018b30d573/nova-cell1-novncproxy-novncproxy/0.log" Oct 14 15:44:26 crc kubenswrapper[4945]: I1014 15:44:26.115169 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f22b1bbe-8060-4302-b929-567dc4f33173/nova-metadata-log/0.log" Oct 14 15:44:26 crc kubenswrapper[4945]: I1014 15:44:26.441721 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_524fdd1e-9a42-495f-a692-e4722ab3c4e2/nova-scheduler-scheduler/0.log" Oct 14 15:44:26 crc kubenswrapper[4945]: I1014 15:44:26.498832 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f0fa2603-a8a2-449d-88d2-e35975a18ad1/mysql-bootstrap/0.log" Oct 14 15:44:26 crc kubenswrapper[4945]: I1014 15:44:26.555510 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f22b1bbe-8060-4302-b929-567dc4f33173/nova-metadata-metadata/0.log" Oct 14 15:44:26 crc kubenswrapper[4945]: I1014 15:44:26.844409 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f0fa2603-a8a2-449d-88d2-e35975a18ad1/mysql-bootstrap/0.log" Oct 14 15:44:26 crc kubenswrapper[4945]: I1014 15:44:26.856555 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f0fa2603-a8a2-449d-88d2-e35975a18ad1/galera/0.log" Oct 14 15:44:27 crc kubenswrapper[4945]: I1014 15:44:27.017912 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_afed9c20-1964-44dd-a849-af9fc0f97e86/mysql-bootstrap/0.log" Oct 14 15:44:27 crc kubenswrapper[4945]: I1014 15:44:27.198650 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_afed9c20-1964-44dd-a849-af9fc0f97e86/mysql-bootstrap/0.log" Oct 14 15:44:27 crc kubenswrapper[4945]: I1014 15:44:27.233018 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_473a83ad-e7a6-4b5b-899e-b64ef41465c9/openstackclient/0.log" Oct 14 15:44:27 crc kubenswrapper[4945]: I1014 15:44:27.242205 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_afed9c20-1964-44dd-a849-af9fc0f97e86/galera/0.log" Oct 14 15:44:27 crc kubenswrapper[4945]: I1014 15:44:27.453996 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6v9cw_9d0edbee-18bd-41ac-880b-526477ef54d3/ovn-controller/0.log" Oct 14 15:44:27 crc kubenswrapper[4945]: I1014 15:44:27.522079 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-krh6j_80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4/openstack-network-exporter/0.log" Oct 14 15:44:27 crc kubenswrapper[4945]: I1014 15:44:27.699116 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pq8xf_2f495edd-59f6-4a56-9356-496cda4ae47a/ovsdb-server-init/0.log" Oct 14 15:44:27 crc kubenswrapper[4945]: I1014 15:44:27.895412 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pq8xf_2f495edd-59f6-4a56-9356-496cda4ae47a/ovs-vswitchd/0.log" Oct 14 15:44:27 crc kubenswrapper[4945]: I1014 15:44:27.906888 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pq8xf_2f495edd-59f6-4a56-9356-496cda4ae47a/ovsdb-server/0.log" Oct 14 15:44:27 crc kubenswrapper[4945]: I1014 15:44:27.978229 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pq8xf_2f495edd-59f6-4a56-9356-496cda4ae47a/ovsdb-server-init/0.log" Oct 14 15:44:28 crc kubenswrapper[4945]: I1014 15:44:28.156675 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31/openstack-network-exporter/0.log" Oct 14 15:44:28 crc kubenswrapper[4945]: I1014 15:44:28.172604 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31/ovn-northd/0.log" Oct 14 15:44:28 crc kubenswrapper[4945]: I1014 15:44:28.265406 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e90e1239-2529-43dd-b429-c140543cb579/openstack-network-exporter/0.log" Oct 14 15:44:28 crc kubenswrapper[4945]: I1014 15:44:28.465408 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e90e1239-2529-43dd-b429-c140543cb579/ovsdbserver-nb/0.log" Oct 14 15:44:28 crc kubenswrapper[4945]: I1014 15:44:28.468477 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3/openstack-network-exporter/0.log" Oct 14 15:44:28 crc kubenswrapper[4945]: I1014 15:44:28.518130 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3/ovsdbserver-sb/0.log" Oct 14 15:44:28 crc kubenswrapper[4945]: I1014 15:44:28.765173 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-bd8df944d-q4z6c_396db799-75e0-4156-a97d-472db3c42729/placement-api/0.log" Oct 14 15:44:28 crc kubenswrapper[4945]: I1014 15:44:28.805520 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-bd8df944d-q4z6c_396db799-75e0-4156-a97d-472db3c42729/placement-log/0.log" Oct 14 15:44:28 crc kubenswrapper[4945]: I1014 15:44:28.954912 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b52be853-0c59-48be-b4d0-3b2023310050/setup-container/0.log" Oct 14 15:44:29 crc kubenswrapper[4945]: I1014 15:44:29.153785 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b52be853-0c59-48be-b4d0-3b2023310050/setup-container/0.log" Oct 14 15:44:29 crc kubenswrapper[4945]: I1014 15:44:29.246210 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b52be853-0c59-48be-b4d0-3b2023310050/rabbitmq/0.log" Oct 14 15:44:29 crc kubenswrapper[4945]: I1014 15:44:29.275674 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_40cdc335-20ee-467f-bc15-4c5867d0fd2a/setup-container/0.log" Oct 14 15:44:29 crc kubenswrapper[4945]: I1014 15:44:29.430526 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_40cdc335-20ee-467f-bc15-4c5867d0fd2a/setup-container/0.log" Oct 14 15:44:29 crc kubenswrapper[4945]: I1014 15:44:29.559989 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_40cdc335-20ee-467f-bc15-4c5867d0fd2a/rabbitmq/0.log" Oct 14 15:44:29 crc kubenswrapper[4945]: I1014 15:44:29.576034 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-86b568845c-dgcjh_047520e2-5898-42bd-85bf-dce4063f08f4/proxy-httpd/0.log" Oct 14 15:44:29 crc kubenswrapper[4945]: I1014 15:44:29.655932 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-86b568845c-dgcjh_047520e2-5898-42bd-85bf-dce4063f08f4/proxy-server/0.log" Oct 14 15:44:29 crc kubenswrapper[4945]: I1014 15:44:29.798861 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-h5b5w_787ea529-5875-4e6c-8a3e-5b4fbf019fa8/swift-ring-rebalance/0.log" Oct 14 15:44:29 crc kubenswrapper[4945]: I1014 15:44:29.886447 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/account-auditor/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.029607 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/account-reaper/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.059655 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/account-replicator/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.081496 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/account-server/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.104545 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/container-auditor/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.247777 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/container-replicator/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.295703 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/container-server/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.330424 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/container-updater/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.331983 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/object-auditor/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.692594 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/object-expirer/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.697471 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/object-replicator/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.763818 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/object-updater/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.813772 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/object-server/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.950150 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/swift-recon-cron/0.log" Oct 14 15:44:30 crc kubenswrapper[4945]: I1014 15:44:30.964014 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/rsync/0.log" Oct 14 15:44:34 crc kubenswrapper[4945]: I1014 15:44:34.473246 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_131ae7de-5f52-4fd7-8789-d61f8a3a9c46/memcached/0.log" Oct 14 15:44:38 crc kubenswrapper[4945]: I1014 15:44:38.763870 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:44:38 crc kubenswrapper[4945]: E1014 15:44:38.764690 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:44:45 crc kubenswrapper[4945]: I1014 15:44:45.745590 4945 scope.go:117] "RemoveContainer" containerID="a041f37d48bb8115b6436162ff26bb850f219513d141dc56dde2b6d342b6a886" Oct 14 15:44:45 crc kubenswrapper[4945]: I1014 15:44:45.775116 4945 scope.go:117] "RemoveContainer" containerID="ddd5cd09cc3e25634705a0e23971e72527bc2353839ffbf5969d4617e8e393b9" Oct 14 15:44:45 crc kubenswrapper[4945]: I1014 15:44:45.825509 4945 scope.go:117] "RemoveContainer" containerID="4d9f58b702d01e9dc96b268b5d5e40d9d58a217ce6c8016419c5bd675cb251e3" Oct 14 15:44:45 crc kubenswrapper[4945]: I1014 15:44:45.880397 4945 scope.go:117] "RemoveContainer" containerID="2c12bdad5e23d70aad90bbf19c2d017a97c95f3c679426d9a0ff06b25caa0089" Oct 14 15:44:45 crc kubenswrapper[4945]: I1014 15:44:45.933925 4945 scope.go:117] "RemoveContainer" containerID="3b038afb3f1c6faadf96de859d0a6d8f9c7dca8a10fee57f405143963947a5a6" Oct 14 15:44:46 crc kubenswrapper[4945]: I1014 15:44:46.005817 4945 scope.go:117] "RemoveContainer" containerID="60be43dbea963d67d0a7034c4d6b7f50b01a064be46324c64f49a9d21c998559" Oct 14 15:44:46 crc kubenswrapper[4945]: I1014 15:44:46.040703 4945 scope.go:117] "RemoveContainer" containerID="1ed91e4e41516acc7e6b8f9ad985d95b34f69a93773e12d062643de0ef1e2d41" Oct 14 15:44:46 crc kubenswrapper[4945]: I1014 15:44:46.127321 4945 scope.go:117] "RemoveContainer" containerID="d735d29138108f68908aa7b5a4bcc6e8783d56a23a5935ab20e297cf863d5279" Oct 14 15:44:52 crc kubenswrapper[4945]: I1014 15:44:52.043059 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-b7scm"] Oct 14 15:44:52 crc kubenswrapper[4945]: I1014 15:44:52.050600 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-b7scm"] Oct 14 15:44:52 crc kubenswrapper[4945]: I1014 15:44:52.774761 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63c34001-2d18-4f96-a9e5-af05eab7f11b" path="/var/lib/kubelet/pods/63c34001-2d18-4f96-a9e5-af05eab7f11b/volumes" Oct 14 15:44:53 crc kubenswrapper[4945]: I1014 15:44:53.158646 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-xlwn6_f5ac9c23-4d14-460d-9255-16d1865e9aa1/kube-rbac-proxy/0.log" Oct 14 15:44:53 crc kubenswrapper[4945]: I1014 15:44:53.208006 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-xlwn6_f5ac9c23-4d14-460d-9255-16d1865e9aa1/manager/0.log" Oct 14 15:44:53 crc kubenswrapper[4945]: I1014 15:44:53.404783 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-8gm4x_3ae6c6f4-2d25-420b-8149-a976de545cd6/kube-rbac-proxy/0.log" Oct 14 15:44:53 crc kubenswrapper[4945]: I1014 15:44:53.438895 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-8gm4x_3ae6c6f4-2d25-420b-8149-a976de545cd6/manager/0.log" Oct 14 15:44:53 crc kubenswrapper[4945]: I1014 15:44:53.553665 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-gf66t_64596f58-bc5a-4cdf-80d4-22002c0c9575/manager/0.log" Oct 14 15:44:53 crc kubenswrapper[4945]: I1014 15:44:53.558968 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-gf66t_64596f58-bc5a-4cdf-80d4-22002c0c9575/kube-rbac-proxy/0.log" Oct 14 15:44:53 crc kubenswrapper[4945]: I1014 15:44:53.664957 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/util/0.log" Oct 14 15:44:53 crc kubenswrapper[4945]: I1014 15:44:53.763483 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:44:53 crc kubenswrapper[4945]: E1014 15:44:53.763957 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:44:53 crc kubenswrapper[4945]: I1014 15:44:53.820562 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/util/0.log" Oct 14 15:44:53 crc kubenswrapper[4945]: I1014 15:44:53.834013 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/pull/0.log" Oct 14 15:44:53 crc kubenswrapper[4945]: I1014 15:44:53.857297 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/pull/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.006031 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/util/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.030726 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fwjnr"] Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.039703 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/pull/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.040467 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fwjnr"] Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.060801 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/extract/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.220258 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-6fg2j_86d698df-b089-498a-98e5-65c196aabcc3/kube-rbac-proxy/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.276595 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-6fg2j_86d698df-b089-498a-98e5-65c196aabcc3/manager/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.293233 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-v9hhr_0a17b021-e8aa-4f88-b150-596885e48ab1/kube-rbac-proxy/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.407799 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-v9hhr_0a17b021-e8aa-4f88-b150-596885e48ab1/manager/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.448233 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-wchqt_80c510fc-48a8-44c1-9927-c9b0ce934410/kube-rbac-proxy/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.479823 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-wchqt_80c510fc-48a8-44c1-9927-c9b0ce934410/manager/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.677310 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-pk25j_41d187ec-0816-4800-9d76-9d7a3f364391/kube-rbac-proxy/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.773494 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10a43a64-535b-4cfe-be95-3f6f06d8eb98" path="/var/lib/kubelet/pods/10a43a64-535b-4cfe-be95-3f6f06d8eb98/volumes" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.838768 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-pk25j_41d187ec-0816-4800-9d76-9d7a3f364391/manager/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.893262 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-76ccf76b65-p9trl_20fbc887-9299-460d-9890-f9d7e0235118/kube-rbac-proxy/0.log" Oct 14 15:44:54 crc kubenswrapper[4945]: I1014 15:44:54.962320 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-76ccf76b65-p9trl_20fbc887-9299-460d-9890-f9d7e0235118/manager/0.log" Oct 14 15:44:55 crc kubenswrapper[4945]: I1014 15:44:55.070180 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-fqb29_97138e7c-7207-4c47-898f-f71625e0e49b/kube-rbac-proxy/0.log" Oct 14 15:44:55 crc kubenswrapper[4945]: I1014 15:44:55.168525 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-fqb29_97138e7c-7207-4c47-898f-f71625e0e49b/manager/0.log" Oct 14 15:44:55 crc kubenswrapper[4945]: I1014 15:44:55.237463 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-jcrs7_2a652464-b536-4402-a883-f52390e0327b/kube-rbac-proxy/0.log" Oct 14 15:44:55 crc kubenswrapper[4945]: I1014 15:44:55.333799 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-jcrs7_2a652464-b536-4402-a883-f52390e0327b/manager/0.log" Oct 14 15:44:55 crc kubenswrapper[4945]: I1014 15:44:55.432048 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-h2p64_d39181a7-852e-402c-9912-7c81a0952e90/kube-rbac-proxy/0.log" Oct 14 15:44:55 crc kubenswrapper[4945]: I1014 15:44:55.514117 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-h2p64_d39181a7-852e-402c-9912-7c81a0952e90/manager/0.log" Oct 14 15:44:55 crc kubenswrapper[4945]: I1014 15:44:55.602991 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-qx6hl_d10e063a-f99e-4fc8-b539-224bbf23f466/kube-rbac-proxy/0.log" Oct 14 15:44:55 crc kubenswrapper[4945]: I1014 15:44:55.677828 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-qx6hl_d10e063a-f99e-4fc8-b539-224bbf23f466/manager/0.log" Oct 14 15:44:55 crc kubenswrapper[4945]: I1014 15:44:55.784722 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-fl79b_b4e68062-2195-43f9-a6ae-a58d12bd7ce2/kube-rbac-proxy/0.log" Oct 14 15:44:55 crc kubenswrapper[4945]: I1014 15:44:55.910770 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-fl79b_b4e68062-2195-43f9-a6ae-a58d12bd7ce2/manager/0.log" Oct 14 15:44:55 crc kubenswrapper[4945]: I1014 15:44:55.971703 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-vspm4_4af90a70-780d-4d6f-9fc5-6a384002a606/kube-rbac-proxy/0.log" Oct 14 15:44:55 crc kubenswrapper[4945]: I1014 15:44:55.996798 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-vspm4_4af90a70-780d-4d6f-9fc5-6a384002a606/manager/0.log" Oct 14 15:44:56 crc kubenswrapper[4945]: I1014 15:44:56.144246 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w_669db3c1-3b7e-4bbd-980e-93fdea61a1cc/kube-rbac-proxy/0.log" Oct 14 15:44:56 crc kubenswrapper[4945]: I1014 15:44:56.229729 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w_669db3c1-3b7e-4bbd-980e-93fdea61a1cc/manager/0.log" Oct 14 15:44:56 crc kubenswrapper[4945]: I1014 15:44:56.400113 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-b9d6d9595-9zj7r_26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad/kube-rbac-proxy/0.log" Oct 14 15:44:56 crc kubenswrapper[4945]: I1014 15:44:56.476899 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-674f597f57-cqpn6_795c3628-2ec0-4846-b345-3019636282e9/kube-rbac-proxy/0.log" Oct 14 15:44:56 crc kubenswrapper[4945]: I1014 15:44:56.813723 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-f9h5q_4b5016e5-1067-4b63-9db4-f5257354538d/registry-server/0.log" Oct 14 15:44:56 crc kubenswrapper[4945]: I1014 15:44:56.881590 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-674f597f57-cqpn6_795c3628-2ec0-4846-b345-3019636282e9/operator/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.009911 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-r6tvw_37b893c7-b6a7-4a19-bc3e-b251deacdd27/kube-rbac-proxy/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.164706 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-r6tvw_37b893c7-b6a7-4a19-bc3e-b251deacdd27/manager/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.275812 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-58vsj_01b34d70-1617-4e4d-81ab-8cada4bb1edf/kube-rbac-proxy/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.280387 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-b9d6d9595-9zj7r_26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad/manager/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.365964 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-58vsj_01b34d70-1617-4e4d-81ab-8cada4bb1edf/manager/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.394538 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q_f23ed77e-a951-4003-bbc2-a2e06431533e/operator/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.484918 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-48x4h_e3193e9c-898b-456d-a825-ea8176568803/kube-rbac-proxy/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.573996 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-48x4h_e3193e9c-898b-456d-a825-ea8176568803/manager/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.645113 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-48769_50453e84-c640-4a80-ac7d-0fe2ca24acfd/kube-rbac-proxy/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.732777 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-48769_50453e84-c640-4a80-ac7d-0fe2ca24acfd/manager/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.820763 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-kkqht_2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c/kube-rbac-proxy/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.833402 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-kkqht_2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c/manager/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.927528 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-sjvtx_2c954bb2-bf12-491f-92e8-637e0eb13ad5/manager/0.log" Oct 14 15:44:57 crc kubenswrapper[4945]: I1014 15:44:57.951514 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-sjvtx_2c954bb2-bf12-491f-92e8-637e0eb13ad5/kube-rbac-proxy/0.log" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.142783 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g"] Oct 14 15:45:00 crc kubenswrapper[4945]: E1014 15:45:00.143602 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="195282c5-c050-4a90-b4cb-97f612fa3286" containerName="container-00" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.143622 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="195282c5-c050-4a90-b4cb-97f612fa3286" containerName="container-00" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.143971 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="195282c5-c050-4a90-b4cb-97f612fa3286" containerName="container-00" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.144804 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.147103 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.147131 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.157985 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g"] Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.274251 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62da2727-3127-4fba-a22a-60ea1ee24191-config-volume\") pod \"collect-profiles-29340945-dw44g\" (UID: \"62da2727-3127-4fba-a22a-60ea1ee24191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.274560 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cmnc\" (UniqueName: \"kubernetes.io/projected/62da2727-3127-4fba-a22a-60ea1ee24191-kube-api-access-4cmnc\") pod \"collect-profiles-29340945-dw44g\" (UID: \"62da2727-3127-4fba-a22a-60ea1ee24191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.274632 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62da2727-3127-4fba-a22a-60ea1ee24191-secret-volume\") pod \"collect-profiles-29340945-dw44g\" (UID: \"62da2727-3127-4fba-a22a-60ea1ee24191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.376460 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62da2727-3127-4fba-a22a-60ea1ee24191-config-volume\") pod \"collect-profiles-29340945-dw44g\" (UID: \"62da2727-3127-4fba-a22a-60ea1ee24191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.376614 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cmnc\" (UniqueName: \"kubernetes.io/projected/62da2727-3127-4fba-a22a-60ea1ee24191-kube-api-access-4cmnc\") pod \"collect-profiles-29340945-dw44g\" (UID: \"62da2727-3127-4fba-a22a-60ea1ee24191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.376649 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62da2727-3127-4fba-a22a-60ea1ee24191-secret-volume\") pod \"collect-profiles-29340945-dw44g\" (UID: \"62da2727-3127-4fba-a22a-60ea1ee24191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.377603 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62da2727-3127-4fba-a22a-60ea1ee24191-config-volume\") pod \"collect-profiles-29340945-dw44g\" (UID: \"62da2727-3127-4fba-a22a-60ea1ee24191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.390622 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62da2727-3127-4fba-a22a-60ea1ee24191-secret-volume\") pod \"collect-profiles-29340945-dw44g\" (UID: \"62da2727-3127-4fba-a22a-60ea1ee24191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.395944 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cmnc\" (UniqueName: \"kubernetes.io/projected/62da2727-3127-4fba-a22a-60ea1ee24191-kube-api-access-4cmnc\") pod \"collect-profiles-29340945-dw44g\" (UID: \"62da2727-3127-4fba-a22a-60ea1ee24191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.465324 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:00 crc kubenswrapper[4945]: I1014 15:45:00.930818 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g"] Oct 14 15:45:01 crc kubenswrapper[4945]: I1014 15:45:01.024186 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" event={"ID":"62da2727-3127-4fba-a22a-60ea1ee24191","Type":"ContainerStarted","Data":"8e07705c1619af0e877804bb87023a9afc9024f33dd0d5cef8f327f111f51fa4"} Oct 14 15:45:02 crc kubenswrapper[4945]: I1014 15:45:02.034081 4945 generic.go:334] "Generic (PLEG): container finished" podID="62da2727-3127-4fba-a22a-60ea1ee24191" containerID="c6844e414c5bee87b2468fbfb04d28247f3d09d1defcfcd1dc3465928862ab94" exitCode=0 Oct 14 15:45:02 crc kubenswrapper[4945]: I1014 15:45:02.034201 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" event={"ID":"62da2727-3127-4fba-a22a-60ea1ee24191","Type":"ContainerDied","Data":"c6844e414c5bee87b2468fbfb04d28247f3d09d1defcfcd1dc3465928862ab94"} Oct 14 15:45:03 crc kubenswrapper[4945]: I1014 15:45:03.401481 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:03 crc kubenswrapper[4945]: I1014 15:45:03.575846 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62da2727-3127-4fba-a22a-60ea1ee24191-config-volume\") pod \"62da2727-3127-4fba-a22a-60ea1ee24191\" (UID: \"62da2727-3127-4fba-a22a-60ea1ee24191\") " Oct 14 15:45:03 crc kubenswrapper[4945]: I1014 15:45:03.576033 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cmnc\" (UniqueName: \"kubernetes.io/projected/62da2727-3127-4fba-a22a-60ea1ee24191-kube-api-access-4cmnc\") pod \"62da2727-3127-4fba-a22a-60ea1ee24191\" (UID: \"62da2727-3127-4fba-a22a-60ea1ee24191\") " Oct 14 15:45:03 crc kubenswrapper[4945]: I1014 15:45:03.576176 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62da2727-3127-4fba-a22a-60ea1ee24191-secret-volume\") pod \"62da2727-3127-4fba-a22a-60ea1ee24191\" (UID: \"62da2727-3127-4fba-a22a-60ea1ee24191\") " Oct 14 15:45:03 crc kubenswrapper[4945]: I1014 15:45:03.576770 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62da2727-3127-4fba-a22a-60ea1ee24191-config-volume" (OuterVolumeSpecName: "config-volume") pod "62da2727-3127-4fba-a22a-60ea1ee24191" (UID: "62da2727-3127-4fba-a22a-60ea1ee24191"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 15:45:03 crc kubenswrapper[4945]: I1014 15:45:03.582627 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62da2727-3127-4fba-a22a-60ea1ee24191-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "62da2727-3127-4fba-a22a-60ea1ee24191" (UID: "62da2727-3127-4fba-a22a-60ea1ee24191"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 15:45:03 crc kubenswrapper[4945]: I1014 15:45:03.583084 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62da2727-3127-4fba-a22a-60ea1ee24191-kube-api-access-4cmnc" (OuterVolumeSpecName: "kube-api-access-4cmnc") pod "62da2727-3127-4fba-a22a-60ea1ee24191" (UID: "62da2727-3127-4fba-a22a-60ea1ee24191"). InnerVolumeSpecName "kube-api-access-4cmnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:45:03 crc kubenswrapper[4945]: I1014 15:45:03.678456 4945 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62da2727-3127-4fba-a22a-60ea1ee24191-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 15:45:03 crc kubenswrapper[4945]: I1014 15:45:03.678503 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cmnc\" (UniqueName: \"kubernetes.io/projected/62da2727-3127-4fba-a22a-60ea1ee24191-kube-api-access-4cmnc\") on node \"crc\" DevicePath \"\"" Oct 14 15:45:03 crc kubenswrapper[4945]: I1014 15:45:03.678519 4945 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62da2727-3127-4fba-a22a-60ea1ee24191-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 15:45:04 crc kubenswrapper[4945]: I1014 15:45:04.059580 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" event={"ID":"62da2727-3127-4fba-a22a-60ea1ee24191","Type":"ContainerDied","Data":"8e07705c1619af0e877804bb87023a9afc9024f33dd0d5cef8f327f111f51fa4"} Oct 14 15:45:04 crc kubenswrapper[4945]: I1014 15:45:04.059625 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e07705c1619af0e877804bb87023a9afc9024f33dd0d5cef8f327f111f51fa4" Oct 14 15:45:04 crc kubenswrapper[4945]: I1014 15:45:04.059671 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340945-dw44g" Oct 14 15:45:05 crc kubenswrapper[4945]: I1014 15:45:05.767468 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:45:05 crc kubenswrapper[4945]: E1014 15:45:05.768679 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:45:13 crc kubenswrapper[4945]: I1014 15:45:13.027224 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jphx9_b47bfeba-2bc6-4c04-bd5a-a02b10ef8478/control-plane-machine-set-operator/0.log" Oct 14 15:45:13 crc kubenswrapper[4945]: I1014 15:45:13.168854 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fkbmp_932ab187-513c-4d0f-9c6f-f986ac444f78/kube-rbac-proxy/0.log" Oct 14 15:45:13 crc kubenswrapper[4945]: I1014 15:45:13.178355 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fkbmp_932ab187-513c-4d0f-9c6f-f986ac444f78/machine-api-operator/0.log" Oct 14 15:45:16 crc kubenswrapper[4945]: I1014 15:45:16.763555 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:45:16 crc kubenswrapper[4945]: E1014 15:45:16.764267 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:45:24 crc kubenswrapper[4945]: I1014 15:45:24.176484 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-jmvq2_6a2a312b-2002-4494-b432-efb6060b5a89/cert-manager-controller/0.log" Oct 14 15:45:24 crc kubenswrapper[4945]: I1014 15:45:24.348146 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-kw8kj_2385f0e3-9e8d-4971-95e6-08dbf4af170c/cert-manager-cainjector/0.log" Oct 14 15:45:24 crc kubenswrapper[4945]: I1014 15:45:24.425574 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-j5ssh_a59d5293-e44c-459d-9875-137a289aa9d8/cert-manager-webhook/0.log" Oct 14 15:45:29 crc kubenswrapper[4945]: I1014 15:45:29.764130 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:45:29 crc kubenswrapper[4945]: E1014 15:45:29.765040 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:45:35 crc kubenswrapper[4945]: I1014 15:45:35.572322 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-27p5l_f481cdcc-3f50-4aba-a28a-500e4b19bcff/nmstate-console-plugin/0.log" Oct 14 15:45:35 crc kubenswrapper[4945]: I1014 15:45:35.749698 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-nx8z6_93cf1921-8abd-42e0-8f80-aa5023f63c96/nmstate-handler/0.log" Oct 14 15:45:35 crc kubenswrapper[4945]: I1014 15:45:35.779991 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-dh8w2_97cf21fd-90e2-4591-8563-66afb4145fa4/kube-rbac-proxy/0.log" Oct 14 15:45:35 crc kubenswrapper[4945]: I1014 15:45:35.832590 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-dh8w2_97cf21fd-90e2-4591-8563-66afb4145fa4/nmstate-metrics/0.log" Oct 14 15:45:35 crc kubenswrapper[4945]: I1014 15:45:35.986236 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-pz6t2_27598c0b-8118-406d-b0dd-9e47a23343ad/nmstate-operator/0.log" Oct 14 15:45:36 crc kubenswrapper[4945]: I1014 15:45:36.057924 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-p64sf_76343e33-0b0d-4df5-98cf-4d58247d41f7/nmstate-webhook/0.log" Oct 14 15:45:40 crc kubenswrapper[4945]: I1014 15:45:40.770499 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:45:40 crc kubenswrapper[4945]: E1014 15:45:40.772823 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:45:41 crc kubenswrapper[4945]: I1014 15:45:41.054104 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-nvcm2"] Oct 14 15:45:41 crc kubenswrapper[4945]: I1014 15:45:41.064833 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-nvcm2"] Oct 14 15:45:42 crc kubenswrapper[4945]: I1014 15:45:42.793494 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2518d4a-a272-4e4e-8c9f-8884d8e4f612" path="/var/lib/kubelet/pods/c2518d4a-a272-4e4e-8c9f-8884d8e4f612/volumes" Oct 14 15:45:46 crc kubenswrapper[4945]: I1014 15:45:46.352004 4945 scope.go:117] "RemoveContainer" containerID="6312c7bc22838dadc44ac2e540aeffc7a384c584fbd94a114b8eec665aa88881" Oct 14 15:45:46 crc kubenswrapper[4945]: I1014 15:45:46.394453 4945 scope.go:117] "RemoveContainer" containerID="d6c931e09a01890871164155e38170b35605e9e2c037ce86c61d29fada94cc7d" Oct 14 15:45:46 crc kubenswrapper[4945]: I1014 15:45:46.451222 4945 scope.go:117] "RemoveContainer" containerID="b3109b3cfe6b3623e04b24200bae138dd64b7a099ddd410a12734b43af5f84f2" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.024093 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-tvn99_38c1e540-b577-472f-82e3-5079b3823d87/kube-rbac-proxy/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.188619 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-tvn99_38c1e540-b577-472f-82e3-5079b3823d87/controller/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.269039 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-frr-files/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.411893 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-reloader/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.428771 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-frr-files/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.459695 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-metrics/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.487196 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-reloader/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.649776 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-reloader/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.659526 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-frr-files/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.679950 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-metrics/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.704478 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-metrics/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.843255 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-frr-files/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.863503 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-metrics/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.863922 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-reloader/0.log" Oct 14 15:45:49 crc kubenswrapper[4945]: I1014 15:45:49.889171 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/controller/0.log" Oct 14 15:45:50 crc kubenswrapper[4945]: I1014 15:45:50.055502 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/kube-rbac-proxy-frr/0.log" Oct 14 15:45:50 crc kubenswrapper[4945]: I1014 15:45:50.063336 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/frr-metrics/0.log" Oct 14 15:45:50 crc kubenswrapper[4945]: I1014 15:45:50.071823 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/kube-rbac-proxy/0.log" Oct 14 15:45:50 crc kubenswrapper[4945]: I1014 15:45:50.270468 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/reloader/0.log" Oct 14 15:45:50 crc kubenswrapper[4945]: I1014 15:45:50.344731 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-pfbt9_79b86a9d-a49c-428a-b3f9-3e2625ad1abc/frr-k8s-webhook-server/0.log" Oct 14 15:45:50 crc kubenswrapper[4945]: I1014 15:45:50.552009 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9f47f9675-47zvd_740d5a8e-2e4a-4fd1-8412-d5b989734f2f/manager/0.log" Oct 14 15:45:50 crc kubenswrapper[4945]: I1014 15:45:50.814944 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6545bcf4ff-f248w_59c45798-dcb8-4fae-ba47-82a861d97023/webhook-server/0.log" Oct 14 15:45:50 crc kubenswrapper[4945]: I1014 15:45:50.841672 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-lgvg5_9a07810a-86fd-4dd6-a704-037bea9afcab/kube-rbac-proxy/0.log" Oct 14 15:45:50 crc kubenswrapper[4945]: I1014 15:45:50.896388 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/frr/0.log" Oct 14 15:45:51 crc kubenswrapper[4945]: I1014 15:45:51.334864 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-lgvg5_9a07810a-86fd-4dd6-a704-037bea9afcab/speaker/0.log" Oct 14 15:45:53 crc kubenswrapper[4945]: I1014 15:45:53.763165 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:45:53 crc kubenswrapper[4945]: E1014 15:45:53.764073 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:46:02 crc kubenswrapper[4945]: I1014 15:46:02.035968 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/util/0.log" Oct 14 15:46:02 crc kubenswrapper[4945]: I1014 15:46:02.208707 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/pull/0.log" Oct 14 15:46:02 crc kubenswrapper[4945]: I1014 15:46:02.227117 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/pull/0.log" Oct 14 15:46:02 crc kubenswrapper[4945]: I1014 15:46:02.232627 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/util/0.log" Oct 14 15:46:02 crc kubenswrapper[4945]: I1014 15:46:02.437640 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/extract/0.log" Oct 14 15:46:02 crc kubenswrapper[4945]: I1014 15:46:02.461147 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/pull/0.log" Oct 14 15:46:02 crc kubenswrapper[4945]: I1014 15:46:02.465842 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/util/0.log" Oct 14 15:46:02 crc kubenswrapper[4945]: I1014 15:46:02.624003 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-n87wl_b3811f6a-94ba-4d10-bd4b-873af175eec4/extract-utilities/0.log" Oct 14 15:46:02 crc kubenswrapper[4945]: I1014 15:46:02.827942 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-n87wl_b3811f6a-94ba-4d10-bd4b-873af175eec4/extract-content/0.log" Oct 14 15:46:02 crc kubenswrapper[4945]: I1014 15:46:02.834198 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-n87wl_b3811f6a-94ba-4d10-bd4b-873af175eec4/extract-content/0.log" Oct 14 15:46:02 crc kubenswrapper[4945]: I1014 15:46:02.834531 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-n87wl_b3811f6a-94ba-4d10-bd4b-873af175eec4/extract-utilities/0.log" Oct 14 15:46:03 crc kubenswrapper[4945]: I1014 15:46:03.007039 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-n87wl_b3811f6a-94ba-4d10-bd4b-873af175eec4/extract-content/0.log" Oct 14 15:46:03 crc kubenswrapper[4945]: I1014 15:46:03.033260 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-n87wl_b3811f6a-94ba-4d10-bd4b-873af175eec4/extract-utilities/0.log" Oct 14 15:46:03 crc kubenswrapper[4945]: I1014 15:46:03.241012 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/extract-utilities/0.log" Oct 14 15:46:03 crc kubenswrapper[4945]: I1014 15:46:03.347469 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-n87wl_b3811f6a-94ba-4d10-bd4b-873af175eec4/registry-server/0.log" Oct 14 15:46:03 crc kubenswrapper[4945]: I1014 15:46:03.406805 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/extract-content/0.log" Oct 14 15:46:03 crc kubenswrapper[4945]: I1014 15:46:03.428483 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/extract-utilities/0.log" Oct 14 15:46:03 crc kubenswrapper[4945]: I1014 15:46:03.469563 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/extract-content/0.log" Oct 14 15:46:03 crc kubenswrapper[4945]: I1014 15:46:03.625807 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/extract-content/0.log" Oct 14 15:46:03 crc kubenswrapper[4945]: I1014 15:46:03.670228 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/extract-utilities/0.log" Oct 14 15:46:03 crc kubenswrapper[4945]: I1014 15:46:03.864353 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/util/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.039730 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/util/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.073328 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/pull/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.101749 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/registry-server/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.118553 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/pull/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.248386 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/util/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.257338 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/pull/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.286591 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/extract/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.421406 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hdrgb_1139d59d-b224-467d-849e-af47cba636e6/marketplace-operator/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.475663 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/extract-utilities/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.701259 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/extract-content/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.701511 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/extract-content/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.724968 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/extract-utilities/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.903328 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/extract-content/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.908856 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/extract-utilities/0.log" Oct 14 15:46:04 crc kubenswrapper[4945]: I1014 15:46:04.991572 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/registry-server/0.log" Oct 14 15:46:05 crc kubenswrapper[4945]: I1014 15:46:05.096108 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/extract-utilities/0.log" Oct 14 15:46:05 crc kubenswrapper[4945]: I1014 15:46:05.245981 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/extract-content/0.log" Oct 14 15:46:05 crc kubenswrapper[4945]: I1014 15:46:05.246272 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/extract-content/0.log" Oct 14 15:46:05 crc kubenswrapper[4945]: I1014 15:46:05.265370 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/extract-utilities/0.log" Oct 14 15:46:05 crc kubenswrapper[4945]: I1014 15:46:05.442371 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/extract-utilities/0.log" Oct 14 15:46:05 crc kubenswrapper[4945]: I1014 15:46:05.448215 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/extract-content/0.log" Oct 14 15:46:05 crc kubenswrapper[4945]: I1014 15:46:05.776486 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/registry-server/0.log" Oct 14 15:46:08 crc kubenswrapper[4945]: I1014 15:46:08.762912 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:46:08 crc kubenswrapper[4945]: E1014 15:46:08.763759 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:46:21 crc kubenswrapper[4945]: I1014 15:46:21.762895 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:46:21 crc kubenswrapper[4945]: E1014 15:46:21.763672 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:46:26 crc kubenswrapper[4945]: E1014 15:46:26.175475 4945 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.245:43730->38.129.56.245:44489: write tcp 38.129.56.245:43730->38.129.56.245:44489: write: broken pipe Oct 14 15:46:33 crc kubenswrapper[4945]: E1014 15:46:33.794456 4945 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.129.56.245:54772->38.129.56.245:44489: read tcp 38.129.56.245:54772->38.129.56.245:44489: read: connection reset by peer Oct 14 15:46:34 crc kubenswrapper[4945]: I1014 15:46:34.763140 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:46:34 crc kubenswrapper[4945]: E1014 15:46:34.763659 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:46:37 crc kubenswrapper[4945]: E1014 15:46:37.723628 4945 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.245:46820->38.129.56.245:44489: write tcp 38.129.56.245:46820->38.129.56.245:44489: write: broken pipe Oct 14 15:46:49 crc kubenswrapper[4945]: I1014 15:46:49.763549 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:46:49 crc kubenswrapper[4945]: E1014 15:46:49.764518 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:47:00 crc kubenswrapper[4945]: I1014 15:47:00.768798 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:47:00 crc kubenswrapper[4945]: E1014 15:47:00.769843 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:47:12 crc kubenswrapper[4945]: I1014 15:47:12.763959 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:47:12 crc kubenswrapper[4945]: E1014 15:47:12.765711 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:47:25 crc kubenswrapper[4945]: I1014 15:47:25.762742 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:47:25 crc kubenswrapper[4945]: E1014 15:47:25.763616 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:47:37 crc kubenswrapper[4945]: I1014 15:47:37.411907 4945 generic.go:334] "Generic (PLEG): container finished" podID="4b474496-81ee-4c0f-b783-6550dd51b361" containerID="f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b" exitCode=0 Oct 14 15:47:37 crc kubenswrapper[4945]: I1014 15:47:37.412079 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ffg6/must-gather-sld2n" event={"ID":"4b474496-81ee-4c0f-b783-6550dd51b361","Type":"ContainerDied","Data":"f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b"} Oct 14 15:47:37 crc kubenswrapper[4945]: I1014 15:47:37.413297 4945 scope.go:117] "RemoveContainer" containerID="f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b" Oct 14 15:47:38 crc kubenswrapper[4945]: I1014 15:47:38.307580 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8ffg6_must-gather-sld2n_4b474496-81ee-4c0f-b783-6550dd51b361/gather/0.log" Oct 14 15:47:39 crc kubenswrapper[4945]: I1014 15:47:39.765247 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:47:39 crc kubenswrapper[4945]: E1014 15:47:39.765615 4945 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tx7tr_openshift-machine-config-operator(08049862-1ea7-4efc-b9ed-df61b47d430b)\"" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" Oct 14 15:47:45 crc kubenswrapper[4945]: I1014 15:47:45.870695 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8ffg6/must-gather-sld2n"] Oct 14 15:47:45 crc kubenswrapper[4945]: I1014 15:47:45.871453 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-8ffg6/must-gather-sld2n" podUID="4b474496-81ee-4c0f-b783-6550dd51b361" containerName="copy" containerID="cri-o://cdb6459ec5df1973d3a02d957e0b250055fe0d254683d380f362c70d059bade0" gracePeriod=2 Oct 14 15:47:45 crc kubenswrapper[4945]: I1014 15:47:45.883596 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8ffg6/must-gather-sld2n"] Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.338517 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8ffg6_must-gather-sld2n_4b474496-81ee-4c0f-b783-6550dd51b361/copy/0.log" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.339520 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/must-gather-sld2n" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.473273 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv9c4\" (UniqueName: \"kubernetes.io/projected/4b474496-81ee-4c0f-b783-6550dd51b361-kube-api-access-cv9c4\") pod \"4b474496-81ee-4c0f-b783-6550dd51b361\" (UID: \"4b474496-81ee-4c0f-b783-6550dd51b361\") " Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.473466 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4b474496-81ee-4c0f-b783-6550dd51b361-must-gather-output\") pod \"4b474496-81ee-4c0f-b783-6550dd51b361\" (UID: \"4b474496-81ee-4c0f-b783-6550dd51b361\") " Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.481538 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b474496-81ee-4c0f-b783-6550dd51b361-kube-api-access-cv9c4" (OuterVolumeSpecName: "kube-api-access-cv9c4") pod "4b474496-81ee-4c0f-b783-6550dd51b361" (UID: "4b474496-81ee-4c0f-b783-6550dd51b361"). InnerVolumeSpecName "kube-api-access-cv9c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.493049 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8ffg6_must-gather-sld2n_4b474496-81ee-4c0f-b783-6550dd51b361/copy/0.log" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.493634 4945 generic.go:334] "Generic (PLEG): container finished" podID="4b474496-81ee-4c0f-b783-6550dd51b361" containerID="cdb6459ec5df1973d3a02d957e0b250055fe0d254683d380f362c70d059bade0" exitCode=143 Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.493708 4945 scope.go:117] "RemoveContainer" containerID="cdb6459ec5df1973d3a02d957e0b250055fe0d254683d380f362c70d059bade0" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.493904 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ffg6/must-gather-sld2n" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.569847 4945 scope.go:117] "RemoveContainer" containerID="f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.575984 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv9c4\" (UniqueName: \"kubernetes.io/projected/4b474496-81ee-4c0f-b783-6550dd51b361-kube-api-access-cv9c4\") on node \"crc\" DevicePath \"\"" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.632560 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b474496-81ee-4c0f-b783-6550dd51b361-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "4b474496-81ee-4c0f-b783-6550dd51b361" (UID: "4b474496-81ee-4c0f-b783-6550dd51b361"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.670294 4945 scope.go:117] "RemoveContainer" containerID="cdb6459ec5df1973d3a02d957e0b250055fe0d254683d380f362c70d059bade0" Oct 14 15:47:46 crc kubenswrapper[4945]: E1014 15:47:46.670779 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdb6459ec5df1973d3a02d957e0b250055fe0d254683d380f362c70d059bade0\": container with ID starting with cdb6459ec5df1973d3a02d957e0b250055fe0d254683d380f362c70d059bade0 not found: ID does not exist" containerID="cdb6459ec5df1973d3a02d957e0b250055fe0d254683d380f362c70d059bade0" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.671162 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdb6459ec5df1973d3a02d957e0b250055fe0d254683d380f362c70d059bade0"} err="failed to get container status \"cdb6459ec5df1973d3a02d957e0b250055fe0d254683d380f362c70d059bade0\": rpc error: code = NotFound desc = could not find container \"cdb6459ec5df1973d3a02d957e0b250055fe0d254683d380f362c70d059bade0\": container with ID starting with cdb6459ec5df1973d3a02d957e0b250055fe0d254683d380f362c70d059bade0 not found: ID does not exist" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.671196 4945 scope.go:117] "RemoveContainer" containerID="f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b" Oct 14 15:47:46 crc kubenswrapper[4945]: E1014 15:47:46.671430 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b\": container with ID starting with f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b not found: ID does not exist" containerID="f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.671460 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b"} err="failed to get container status \"f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b\": rpc error: code = NotFound desc = could not find container \"f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b\": container with ID starting with f375d20dfa61a4faa83e602ad7c18ae2073c39069e173a98b4190de25a0b755b not found: ID does not exist" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.677988 4945 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4b474496-81ee-4c0f-b783-6550dd51b361-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 14 15:47:46 crc kubenswrapper[4945]: I1014 15:47:46.774122 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b474496-81ee-4c0f-b783-6550dd51b361" path="/var/lib/kubelet/pods/4b474496-81ee-4c0f-b783-6550dd51b361/volumes" Oct 14 15:47:53 crc kubenswrapper[4945]: I1014 15:47:53.763577 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:47:54 crc kubenswrapper[4945]: I1014 15:47:54.567183 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerStarted","Data":"0a877f7512be56a61764de5bfd6cb2d36289868e677eb3453925231a2256b2a9"} Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.515692 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6v2cl/must-gather-tw6jh"] Oct 14 15:48:46 crc kubenswrapper[4945]: E1014 15:48:46.516610 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b474496-81ee-4c0f-b783-6550dd51b361" containerName="gather" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.516624 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b474496-81ee-4c0f-b783-6550dd51b361" containerName="gather" Oct 14 15:48:46 crc kubenswrapper[4945]: E1014 15:48:46.516643 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b474496-81ee-4c0f-b783-6550dd51b361" containerName="copy" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.516649 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b474496-81ee-4c0f-b783-6550dd51b361" containerName="copy" Oct 14 15:48:46 crc kubenswrapper[4945]: E1014 15:48:46.516668 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62da2727-3127-4fba-a22a-60ea1ee24191" containerName="collect-profiles" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.516674 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="62da2727-3127-4fba-a22a-60ea1ee24191" containerName="collect-profiles" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.516927 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b474496-81ee-4c0f-b783-6550dd51b361" containerName="copy" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.516958 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="62da2727-3127-4fba-a22a-60ea1ee24191" containerName="collect-profiles" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.516974 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b474496-81ee-4c0f-b783-6550dd51b361" containerName="gather" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.521020 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/must-gather-tw6jh" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.522721 4945 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-6v2cl"/"default-dockercfg-lkgsg" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.524525 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6v2cl"/"kube-root-ca.crt" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.524797 4945 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6v2cl"/"openshift-service-ca.crt" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.544340 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6v2cl/must-gather-tw6jh"] Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.590183 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqdpj\" (UniqueName: \"kubernetes.io/projected/6eaf1e4a-64f3-4cba-a83d-900ba092249c-kube-api-access-kqdpj\") pod \"must-gather-tw6jh\" (UID: \"6eaf1e4a-64f3-4cba-a83d-900ba092249c\") " pod="openshift-must-gather-6v2cl/must-gather-tw6jh" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.590553 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6eaf1e4a-64f3-4cba-a83d-900ba092249c-must-gather-output\") pod \"must-gather-tw6jh\" (UID: \"6eaf1e4a-64f3-4cba-a83d-900ba092249c\") " pod="openshift-must-gather-6v2cl/must-gather-tw6jh" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.692268 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqdpj\" (UniqueName: \"kubernetes.io/projected/6eaf1e4a-64f3-4cba-a83d-900ba092249c-kube-api-access-kqdpj\") pod \"must-gather-tw6jh\" (UID: \"6eaf1e4a-64f3-4cba-a83d-900ba092249c\") " pod="openshift-must-gather-6v2cl/must-gather-tw6jh" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.692445 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6eaf1e4a-64f3-4cba-a83d-900ba092249c-must-gather-output\") pod \"must-gather-tw6jh\" (UID: \"6eaf1e4a-64f3-4cba-a83d-900ba092249c\") " pod="openshift-must-gather-6v2cl/must-gather-tw6jh" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.693038 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6eaf1e4a-64f3-4cba-a83d-900ba092249c-must-gather-output\") pod \"must-gather-tw6jh\" (UID: \"6eaf1e4a-64f3-4cba-a83d-900ba092249c\") " pod="openshift-must-gather-6v2cl/must-gather-tw6jh" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.712636 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqdpj\" (UniqueName: \"kubernetes.io/projected/6eaf1e4a-64f3-4cba-a83d-900ba092249c-kube-api-access-kqdpj\") pod \"must-gather-tw6jh\" (UID: \"6eaf1e4a-64f3-4cba-a83d-900ba092249c\") " pod="openshift-must-gather-6v2cl/must-gather-tw6jh" Oct 14 15:48:46 crc kubenswrapper[4945]: I1014 15:48:46.848999 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/must-gather-tw6jh" Oct 14 15:48:47 crc kubenswrapper[4945]: I1014 15:48:47.471835 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6v2cl/must-gather-tw6jh"] Oct 14 15:48:48 crc kubenswrapper[4945]: I1014 15:48:48.073995 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6v2cl/must-gather-tw6jh" event={"ID":"6eaf1e4a-64f3-4cba-a83d-900ba092249c","Type":"ContainerStarted","Data":"5d62a6359df9872e4c89fbdca62fe2f9cf9a771a978df806b9e03c91ea3dc72e"} Oct 14 15:48:48 crc kubenswrapper[4945]: I1014 15:48:48.074327 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6v2cl/must-gather-tw6jh" event={"ID":"6eaf1e4a-64f3-4cba-a83d-900ba092249c","Type":"ContainerStarted","Data":"6e65a24520ae1bdd6636af8e8c55bc0923d07f1b61baeb2586d875fde80306a1"} Oct 14 15:48:48 crc kubenswrapper[4945]: I1014 15:48:48.074351 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6v2cl/must-gather-tw6jh" event={"ID":"6eaf1e4a-64f3-4cba-a83d-900ba092249c","Type":"ContainerStarted","Data":"02be56e4602d9fe551a84bb38428d65b85e7a50857f220ce15372569649aed51"} Oct 14 15:48:48 crc kubenswrapper[4945]: I1014 15:48:48.099181 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6v2cl/must-gather-tw6jh" podStartSLOduration=2.099162708 podStartE2EDuration="2.099162708s" podCreationTimestamp="2025-10-14 15:48:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:48:48.094021142 +0000 UTC m=+2058.078069540" watchObservedRunningTime="2025-10-14 15:48:48.099162708 +0000 UTC m=+2058.083211076" Oct 14 15:48:51 crc kubenswrapper[4945]: I1014 15:48:51.763892 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6v2cl/crc-debug-4fv7k"] Oct 14 15:48:51 crc kubenswrapper[4945]: I1014 15:48:51.769327 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" Oct 14 15:48:51 crc kubenswrapper[4945]: I1014 15:48:51.898616 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0133f8f-d913-45e2-860a-1927c562b1be-host\") pod \"crc-debug-4fv7k\" (UID: \"f0133f8f-d913-45e2-860a-1927c562b1be\") " pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" Oct 14 15:48:51 crc kubenswrapper[4945]: I1014 15:48:51.899410 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srxhq\" (UniqueName: \"kubernetes.io/projected/f0133f8f-d913-45e2-860a-1927c562b1be-kube-api-access-srxhq\") pod \"crc-debug-4fv7k\" (UID: \"f0133f8f-d913-45e2-860a-1927c562b1be\") " pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" Oct 14 15:48:52 crc kubenswrapper[4945]: I1014 15:48:52.001151 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0133f8f-d913-45e2-860a-1927c562b1be-host\") pod \"crc-debug-4fv7k\" (UID: \"f0133f8f-d913-45e2-860a-1927c562b1be\") " pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" Oct 14 15:48:52 crc kubenswrapper[4945]: I1014 15:48:52.001241 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srxhq\" (UniqueName: \"kubernetes.io/projected/f0133f8f-d913-45e2-860a-1927c562b1be-kube-api-access-srxhq\") pod \"crc-debug-4fv7k\" (UID: \"f0133f8f-d913-45e2-860a-1927c562b1be\") " pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" Oct 14 15:48:52 crc kubenswrapper[4945]: I1014 15:48:52.001353 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0133f8f-d913-45e2-860a-1927c562b1be-host\") pod \"crc-debug-4fv7k\" (UID: \"f0133f8f-d913-45e2-860a-1927c562b1be\") " pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" Oct 14 15:48:52 crc kubenswrapper[4945]: I1014 15:48:52.020947 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srxhq\" (UniqueName: \"kubernetes.io/projected/f0133f8f-d913-45e2-860a-1927c562b1be-kube-api-access-srxhq\") pod \"crc-debug-4fv7k\" (UID: \"f0133f8f-d913-45e2-860a-1927c562b1be\") " pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" Oct 14 15:48:52 crc kubenswrapper[4945]: I1014 15:48:52.098422 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" Oct 14 15:48:52 crc kubenswrapper[4945]: W1014 15:48:52.159250 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0133f8f_d913_45e2_860a_1927c562b1be.slice/crio-013b5d52f42de76a7c77aa60b8e7bf3733bbc958eb01d198365511f86eb9d7f8 WatchSource:0}: Error finding container 013b5d52f42de76a7c77aa60b8e7bf3733bbc958eb01d198365511f86eb9d7f8: Status 404 returned error can't find the container with id 013b5d52f42de76a7c77aa60b8e7bf3733bbc958eb01d198365511f86eb9d7f8 Oct 14 15:48:53 crc kubenswrapper[4945]: I1014 15:48:53.126923 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" event={"ID":"f0133f8f-d913-45e2-860a-1927c562b1be","Type":"ContainerStarted","Data":"5f6b8ce4b6be55070b7c2d5440a57b2a0fa43ec45301718b20af20a034951842"} Oct 14 15:48:53 crc kubenswrapper[4945]: I1014 15:48:53.127536 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" event={"ID":"f0133f8f-d913-45e2-860a-1927c562b1be","Type":"ContainerStarted","Data":"013b5d52f42de76a7c77aa60b8e7bf3733bbc958eb01d198365511f86eb9d7f8"} Oct 14 15:48:53 crc kubenswrapper[4945]: I1014 15:48:53.141175 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" podStartSLOduration=2.141158843 podStartE2EDuration="2.141158843s" podCreationTimestamp="2025-10-14 15:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 15:48:53.139952809 +0000 UTC m=+2063.124001197" watchObservedRunningTime="2025-10-14 15:48:53.141158843 +0000 UTC m=+2063.125207211" Oct 14 15:49:26 crc kubenswrapper[4945]: I1014 15:49:26.412323 4945 generic.go:334] "Generic (PLEG): container finished" podID="f0133f8f-d913-45e2-860a-1927c562b1be" containerID="5f6b8ce4b6be55070b7c2d5440a57b2a0fa43ec45301718b20af20a034951842" exitCode=0 Oct 14 15:49:26 crc kubenswrapper[4945]: I1014 15:49:26.412422 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" event={"ID":"f0133f8f-d913-45e2-860a-1927c562b1be","Type":"ContainerDied","Data":"5f6b8ce4b6be55070b7c2d5440a57b2a0fa43ec45301718b20af20a034951842"} Oct 14 15:49:27 crc kubenswrapper[4945]: I1014 15:49:27.529223 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" Oct 14 15:49:27 crc kubenswrapper[4945]: I1014 15:49:27.559273 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6v2cl/crc-debug-4fv7k"] Oct 14 15:49:27 crc kubenswrapper[4945]: I1014 15:49:27.566331 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6v2cl/crc-debug-4fv7k"] Oct 14 15:49:27 crc kubenswrapper[4945]: I1014 15:49:27.653189 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0133f8f-d913-45e2-860a-1927c562b1be-host\") pod \"f0133f8f-d913-45e2-860a-1927c562b1be\" (UID: \"f0133f8f-d913-45e2-860a-1927c562b1be\") " Oct 14 15:49:27 crc kubenswrapper[4945]: I1014 15:49:27.653491 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srxhq\" (UniqueName: \"kubernetes.io/projected/f0133f8f-d913-45e2-860a-1927c562b1be-kube-api-access-srxhq\") pod \"f0133f8f-d913-45e2-860a-1927c562b1be\" (UID: \"f0133f8f-d913-45e2-860a-1927c562b1be\") " Oct 14 15:49:27 crc kubenswrapper[4945]: I1014 15:49:27.653332 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0133f8f-d913-45e2-860a-1927c562b1be-host" (OuterVolumeSpecName: "host") pod "f0133f8f-d913-45e2-860a-1927c562b1be" (UID: "f0133f8f-d913-45e2-860a-1927c562b1be"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:49:27 crc kubenswrapper[4945]: I1014 15:49:27.654178 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0133f8f-d913-45e2-860a-1927c562b1be-host\") on node \"crc\" DevicePath \"\"" Oct 14 15:49:27 crc kubenswrapper[4945]: I1014 15:49:27.658016 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0133f8f-d913-45e2-860a-1927c562b1be-kube-api-access-srxhq" (OuterVolumeSpecName: "kube-api-access-srxhq") pod "f0133f8f-d913-45e2-860a-1927c562b1be" (UID: "f0133f8f-d913-45e2-860a-1927c562b1be"). InnerVolumeSpecName "kube-api-access-srxhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:49:27 crc kubenswrapper[4945]: I1014 15:49:27.755935 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srxhq\" (UniqueName: \"kubernetes.io/projected/f0133f8f-d913-45e2-860a-1927c562b1be-kube-api-access-srxhq\") on node \"crc\" DevicePath \"\"" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.432511 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="013b5d52f42de76a7c77aa60b8e7bf3733bbc958eb01d198365511f86eb9d7f8" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.432595 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/crc-debug-4fv7k" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.753515 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6v2cl/crc-debug-trxhf"] Oct 14 15:49:28 crc kubenswrapper[4945]: E1014 15:49:28.753959 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0133f8f-d913-45e2-860a-1927c562b1be" containerName="container-00" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.753970 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0133f8f-d913-45e2-860a-1927c562b1be" containerName="container-00" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.754156 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0133f8f-d913-45e2-860a-1927c562b1be" containerName="container-00" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.754864 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/crc-debug-trxhf" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.776896 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0133f8f-d913-45e2-860a-1927c562b1be" path="/var/lib/kubelet/pods/f0133f8f-d913-45e2-860a-1927c562b1be/volumes" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.875108 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx7bv\" (UniqueName: \"kubernetes.io/projected/83a8966c-3784-4d97-9cac-b6a1c4561ad6-kube-api-access-vx7bv\") pod \"crc-debug-trxhf\" (UID: \"83a8966c-3784-4d97-9cac-b6a1c4561ad6\") " pod="openshift-must-gather-6v2cl/crc-debug-trxhf" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.875245 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/83a8966c-3784-4d97-9cac-b6a1c4561ad6-host\") pod \"crc-debug-trxhf\" (UID: \"83a8966c-3784-4d97-9cac-b6a1c4561ad6\") " pod="openshift-must-gather-6v2cl/crc-debug-trxhf" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.976754 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/83a8966c-3784-4d97-9cac-b6a1c4561ad6-host\") pod \"crc-debug-trxhf\" (UID: \"83a8966c-3784-4d97-9cac-b6a1c4561ad6\") " pod="openshift-must-gather-6v2cl/crc-debug-trxhf" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.976909 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx7bv\" (UniqueName: \"kubernetes.io/projected/83a8966c-3784-4d97-9cac-b6a1c4561ad6-kube-api-access-vx7bv\") pod \"crc-debug-trxhf\" (UID: \"83a8966c-3784-4d97-9cac-b6a1c4561ad6\") " pod="openshift-must-gather-6v2cl/crc-debug-trxhf" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.977309 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/83a8966c-3784-4d97-9cac-b6a1c4561ad6-host\") pod \"crc-debug-trxhf\" (UID: \"83a8966c-3784-4d97-9cac-b6a1c4561ad6\") " pod="openshift-must-gather-6v2cl/crc-debug-trxhf" Oct 14 15:49:28 crc kubenswrapper[4945]: I1014 15:49:28.998540 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx7bv\" (UniqueName: \"kubernetes.io/projected/83a8966c-3784-4d97-9cac-b6a1c4561ad6-kube-api-access-vx7bv\") pod \"crc-debug-trxhf\" (UID: \"83a8966c-3784-4d97-9cac-b6a1c4561ad6\") " pod="openshift-must-gather-6v2cl/crc-debug-trxhf" Oct 14 15:49:29 crc kubenswrapper[4945]: I1014 15:49:29.076436 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/crc-debug-trxhf" Oct 14 15:49:29 crc kubenswrapper[4945]: I1014 15:49:29.444040 4945 generic.go:334] "Generic (PLEG): container finished" podID="83a8966c-3784-4d97-9cac-b6a1c4561ad6" containerID="b115cd3b4b655393ebf3a01b757fc7f88de6d21478ef4f62ae6eaf20ec05ade0" exitCode=0 Oct 14 15:49:29 crc kubenswrapper[4945]: I1014 15:49:29.444159 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6v2cl/crc-debug-trxhf" event={"ID":"83a8966c-3784-4d97-9cac-b6a1c4561ad6","Type":"ContainerDied","Data":"b115cd3b4b655393ebf3a01b757fc7f88de6d21478ef4f62ae6eaf20ec05ade0"} Oct 14 15:49:29 crc kubenswrapper[4945]: I1014 15:49:29.444467 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6v2cl/crc-debug-trxhf" event={"ID":"83a8966c-3784-4d97-9cac-b6a1c4561ad6","Type":"ContainerStarted","Data":"0952d34a8b0fe544ed8046ec0396eef638dde6d3d811210c3346511a0db492c4"} Oct 14 15:49:29 crc kubenswrapper[4945]: I1014 15:49:29.836378 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6v2cl/crc-debug-trxhf"] Oct 14 15:49:29 crc kubenswrapper[4945]: I1014 15:49:29.842690 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6v2cl/crc-debug-trxhf"] Oct 14 15:49:30 crc kubenswrapper[4945]: I1014 15:49:30.546551 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/crc-debug-trxhf" Oct 14 15:49:30 crc kubenswrapper[4945]: I1014 15:49:30.603434 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vx7bv\" (UniqueName: \"kubernetes.io/projected/83a8966c-3784-4d97-9cac-b6a1c4561ad6-kube-api-access-vx7bv\") pod \"83a8966c-3784-4d97-9cac-b6a1c4561ad6\" (UID: \"83a8966c-3784-4d97-9cac-b6a1c4561ad6\") " Oct 14 15:49:30 crc kubenswrapper[4945]: I1014 15:49:30.603487 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/83a8966c-3784-4d97-9cac-b6a1c4561ad6-host\") pod \"83a8966c-3784-4d97-9cac-b6a1c4561ad6\" (UID: \"83a8966c-3784-4d97-9cac-b6a1c4561ad6\") " Oct 14 15:49:30 crc kubenswrapper[4945]: I1014 15:49:30.604136 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83a8966c-3784-4d97-9cac-b6a1c4561ad6-host" (OuterVolumeSpecName: "host") pod "83a8966c-3784-4d97-9cac-b6a1c4561ad6" (UID: "83a8966c-3784-4d97-9cac-b6a1c4561ad6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:49:30 crc kubenswrapper[4945]: I1014 15:49:30.609216 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a8966c-3784-4d97-9cac-b6a1c4561ad6-kube-api-access-vx7bv" (OuterVolumeSpecName: "kube-api-access-vx7bv") pod "83a8966c-3784-4d97-9cac-b6a1c4561ad6" (UID: "83a8966c-3784-4d97-9cac-b6a1c4561ad6"). InnerVolumeSpecName "kube-api-access-vx7bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:49:30 crc kubenswrapper[4945]: I1014 15:49:30.706304 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vx7bv\" (UniqueName: \"kubernetes.io/projected/83a8966c-3784-4d97-9cac-b6a1c4561ad6-kube-api-access-vx7bv\") on node \"crc\" DevicePath \"\"" Oct 14 15:49:30 crc kubenswrapper[4945]: I1014 15:49:30.706360 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/83a8966c-3784-4d97-9cac-b6a1c4561ad6-host\") on node \"crc\" DevicePath \"\"" Oct 14 15:49:30 crc kubenswrapper[4945]: I1014 15:49:30.779714 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a8966c-3784-4d97-9cac-b6a1c4561ad6" path="/var/lib/kubelet/pods/83a8966c-3784-4d97-9cac-b6a1c4561ad6/volumes" Oct 14 15:49:30 crc kubenswrapper[4945]: I1014 15:49:30.992795 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6v2cl/crc-debug-2jl9p"] Oct 14 15:49:30 crc kubenswrapper[4945]: E1014 15:49:30.993189 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a8966c-3784-4d97-9cac-b6a1c4561ad6" containerName="container-00" Oct 14 15:49:30 crc kubenswrapper[4945]: I1014 15:49:30.993203 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a8966c-3784-4d97-9cac-b6a1c4561ad6" containerName="container-00" Oct 14 15:49:30 crc kubenswrapper[4945]: I1014 15:49:30.993407 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a8966c-3784-4d97-9cac-b6a1c4561ad6" containerName="container-00" Oct 14 15:49:30 crc kubenswrapper[4945]: I1014 15:49:30.993990 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/crc-debug-2jl9p" Oct 14 15:49:31 crc kubenswrapper[4945]: I1014 15:49:31.112922 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5pg9\" (UniqueName: \"kubernetes.io/projected/7a9000a7-081c-4158-9551-62f3c8feffa8-kube-api-access-t5pg9\") pod \"crc-debug-2jl9p\" (UID: \"7a9000a7-081c-4158-9551-62f3c8feffa8\") " pod="openshift-must-gather-6v2cl/crc-debug-2jl9p" Oct 14 15:49:31 crc kubenswrapper[4945]: I1014 15:49:31.113048 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a9000a7-081c-4158-9551-62f3c8feffa8-host\") pod \"crc-debug-2jl9p\" (UID: \"7a9000a7-081c-4158-9551-62f3c8feffa8\") " pod="openshift-must-gather-6v2cl/crc-debug-2jl9p" Oct 14 15:49:31 crc kubenswrapper[4945]: I1014 15:49:31.214394 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5pg9\" (UniqueName: \"kubernetes.io/projected/7a9000a7-081c-4158-9551-62f3c8feffa8-kube-api-access-t5pg9\") pod \"crc-debug-2jl9p\" (UID: \"7a9000a7-081c-4158-9551-62f3c8feffa8\") " pod="openshift-must-gather-6v2cl/crc-debug-2jl9p" Oct 14 15:49:31 crc kubenswrapper[4945]: I1014 15:49:31.214489 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a9000a7-081c-4158-9551-62f3c8feffa8-host\") pod \"crc-debug-2jl9p\" (UID: \"7a9000a7-081c-4158-9551-62f3c8feffa8\") " pod="openshift-must-gather-6v2cl/crc-debug-2jl9p" Oct 14 15:49:31 crc kubenswrapper[4945]: I1014 15:49:31.214629 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a9000a7-081c-4158-9551-62f3c8feffa8-host\") pod \"crc-debug-2jl9p\" (UID: \"7a9000a7-081c-4158-9551-62f3c8feffa8\") " pod="openshift-must-gather-6v2cl/crc-debug-2jl9p" Oct 14 15:49:31 crc kubenswrapper[4945]: I1014 15:49:31.231800 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5pg9\" (UniqueName: \"kubernetes.io/projected/7a9000a7-081c-4158-9551-62f3c8feffa8-kube-api-access-t5pg9\") pod \"crc-debug-2jl9p\" (UID: \"7a9000a7-081c-4158-9551-62f3c8feffa8\") " pod="openshift-must-gather-6v2cl/crc-debug-2jl9p" Oct 14 15:49:31 crc kubenswrapper[4945]: I1014 15:49:31.311898 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/crc-debug-2jl9p" Oct 14 15:49:31 crc kubenswrapper[4945]: I1014 15:49:31.478017 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6v2cl/crc-debug-2jl9p" event={"ID":"7a9000a7-081c-4158-9551-62f3c8feffa8","Type":"ContainerStarted","Data":"1db934c5060714d268946a7035515cd05b897c92da33cbce7387339e4ae8e94b"} Oct 14 15:49:31 crc kubenswrapper[4945]: I1014 15:49:31.480313 4945 scope.go:117] "RemoveContainer" containerID="b115cd3b4b655393ebf3a01b757fc7f88de6d21478ef4f62ae6eaf20ec05ade0" Oct 14 15:49:31 crc kubenswrapper[4945]: I1014 15:49:31.480524 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/crc-debug-trxhf" Oct 14 15:49:32 crc kubenswrapper[4945]: I1014 15:49:32.490704 4945 generic.go:334] "Generic (PLEG): container finished" podID="7a9000a7-081c-4158-9551-62f3c8feffa8" containerID="a46bce581950f37554b7409265f7d996455292ef9429af3ca4eea7e07c60cd4c" exitCode=0 Oct 14 15:49:32 crc kubenswrapper[4945]: I1014 15:49:32.490773 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6v2cl/crc-debug-2jl9p" event={"ID":"7a9000a7-081c-4158-9551-62f3c8feffa8","Type":"ContainerDied","Data":"a46bce581950f37554b7409265f7d996455292ef9429af3ca4eea7e07c60cd4c"} Oct 14 15:49:32 crc kubenswrapper[4945]: I1014 15:49:32.535352 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6v2cl/crc-debug-2jl9p"] Oct 14 15:49:32 crc kubenswrapper[4945]: I1014 15:49:32.543970 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6v2cl/crc-debug-2jl9p"] Oct 14 15:49:32 crc kubenswrapper[4945]: E1014 15:49:32.949784 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a8966c_3784_4d97_9cac_b6a1c4561ad6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a8966c_3784_4d97_9cac_b6a1c4561ad6.slice/crio-0952d34a8b0fe544ed8046ec0396eef638dde6d3d811210c3346511a0db492c4\": RecentStats: unable to find data in memory cache]" Oct 14 15:49:33 crc kubenswrapper[4945]: I1014 15:49:33.599631 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/crc-debug-2jl9p" Oct 14 15:49:33 crc kubenswrapper[4945]: I1014 15:49:33.662276 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5pg9\" (UniqueName: \"kubernetes.io/projected/7a9000a7-081c-4158-9551-62f3c8feffa8-kube-api-access-t5pg9\") pod \"7a9000a7-081c-4158-9551-62f3c8feffa8\" (UID: \"7a9000a7-081c-4158-9551-62f3c8feffa8\") " Oct 14 15:49:33 crc kubenswrapper[4945]: I1014 15:49:33.662427 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a9000a7-081c-4158-9551-62f3c8feffa8-host\") pod \"7a9000a7-081c-4158-9551-62f3c8feffa8\" (UID: \"7a9000a7-081c-4158-9551-62f3c8feffa8\") " Oct 14 15:49:33 crc kubenswrapper[4945]: I1014 15:49:33.662630 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a9000a7-081c-4158-9551-62f3c8feffa8-host" (OuterVolumeSpecName: "host") pod "7a9000a7-081c-4158-9551-62f3c8feffa8" (UID: "7a9000a7-081c-4158-9551-62f3c8feffa8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 15:49:33 crc kubenswrapper[4945]: I1014 15:49:33.662855 4945 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a9000a7-081c-4158-9551-62f3c8feffa8-host\") on node \"crc\" DevicePath \"\"" Oct 14 15:49:33 crc kubenswrapper[4945]: I1014 15:49:33.673032 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a9000a7-081c-4158-9551-62f3c8feffa8-kube-api-access-t5pg9" (OuterVolumeSpecName: "kube-api-access-t5pg9") pod "7a9000a7-081c-4158-9551-62f3c8feffa8" (UID: "7a9000a7-081c-4158-9551-62f3c8feffa8"). InnerVolumeSpecName "kube-api-access-t5pg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:49:33 crc kubenswrapper[4945]: I1014 15:49:33.764314 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5pg9\" (UniqueName: \"kubernetes.io/projected/7a9000a7-081c-4158-9551-62f3c8feffa8-kube-api-access-t5pg9\") on node \"crc\" DevicePath \"\"" Oct 14 15:49:34 crc kubenswrapper[4945]: I1014 15:49:34.513384 4945 scope.go:117] "RemoveContainer" containerID="a46bce581950f37554b7409265f7d996455292ef9429af3ca4eea7e07c60cd4c" Oct 14 15:49:34 crc kubenswrapper[4945]: I1014 15:49:34.513506 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/crc-debug-2jl9p" Oct 14 15:49:34 crc kubenswrapper[4945]: I1014 15:49:34.774458 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a9000a7-081c-4158-9551-62f3c8feffa8" path="/var/lib/kubelet/pods/7a9000a7-081c-4158-9551-62f3c8feffa8/volumes" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.004716 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r9phx"] Oct 14 15:49:41 crc kubenswrapper[4945]: E1014 15:49:41.005656 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a9000a7-081c-4158-9551-62f3c8feffa8" containerName="container-00" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.005669 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a9000a7-081c-4158-9551-62f3c8feffa8" containerName="container-00" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.005848 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a9000a7-081c-4158-9551-62f3c8feffa8" containerName="container-00" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.007693 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.031151 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r9phx"] Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.111469 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa27100b-ae33-485e-b88d-3a695cab542a-catalog-content\") pod \"redhat-marketplace-r9phx\" (UID: \"aa27100b-ae33-485e-b88d-3a695cab542a\") " pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.111623 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa27100b-ae33-485e-b88d-3a695cab542a-utilities\") pod \"redhat-marketplace-r9phx\" (UID: \"aa27100b-ae33-485e-b88d-3a695cab542a\") " pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.111649 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9254\" (UniqueName: \"kubernetes.io/projected/aa27100b-ae33-485e-b88d-3a695cab542a-kube-api-access-z9254\") pod \"redhat-marketplace-r9phx\" (UID: \"aa27100b-ae33-485e-b88d-3a695cab542a\") " pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.214248 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa27100b-ae33-485e-b88d-3a695cab542a-catalog-content\") pod \"redhat-marketplace-r9phx\" (UID: \"aa27100b-ae33-485e-b88d-3a695cab542a\") " pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.214374 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa27100b-ae33-485e-b88d-3a695cab542a-utilities\") pod \"redhat-marketplace-r9phx\" (UID: \"aa27100b-ae33-485e-b88d-3a695cab542a\") " pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.214397 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9254\" (UniqueName: \"kubernetes.io/projected/aa27100b-ae33-485e-b88d-3a695cab542a-kube-api-access-z9254\") pod \"redhat-marketplace-r9phx\" (UID: \"aa27100b-ae33-485e-b88d-3a695cab542a\") " pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.214714 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa27100b-ae33-485e-b88d-3a695cab542a-catalog-content\") pod \"redhat-marketplace-r9phx\" (UID: \"aa27100b-ae33-485e-b88d-3a695cab542a\") " pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.214847 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa27100b-ae33-485e-b88d-3a695cab542a-utilities\") pod \"redhat-marketplace-r9phx\" (UID: \"aa27100b-ae33-485e-b88d-3a695cab542a\") " pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.242650 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9254\" (UniqueName: \"kubernetes.io/projected/aa27100b-ae33-485e-b88d-3a695cab542a-kube-api-access-z9254\") pod \"redhat-marketplace-r9phx\" (UID: \"aa27100b-ae33-485e-b88d-3a695cab542a\") " pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.330702 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:41 crc kubenswrapper[4945]: I1014 15:49:41.807269 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r9phx"] Oct 14 15:49:42 crc kubenswrapper[4945]: I1014 15:49:42.583025 4945 generic.go:334] "Generic (PLEG): container finished" podID="aa27100b-ae33-485e-b88d-3a695cab542a" containerID="1e89d26ce41df9312296ac9fd8a26c61f4a2c828d9aefd4d2f36f660945906e4" exitCode=0 Oct 14 15:49:42 crc kubenswrapper[4945]: I1014 15:49:42.583367 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r9phx" event={"ID":"aa27100b-ae33-485e-b88d-3a695cab542a","Type":"ContainerDied","Data":"1e89d26ce41df9312296ac9fd8a26c61f4a2c828d9aefd4d2f36f660945906e4"} Oct 14 15:49:42 crc kubenswrapper[4945]: I1014 15:49:42.583397 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r9phx" event={"ID":"aa27100b-ae33-485e-b88d-3a695cab542a","Type":"ContainerStarted","Data":"4b783fcc116091cae372f5f971a202bb06c6248a150bad93a427ac9998e04acc"} Oct 14 15:49:42 crc kubenswrapper[4945]: I1014 15:49:42.585364 4945 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 15:49:43 crc kubenswrapper[4945]: E1014 15:49:43.203543 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a8966c_3784_4d97_9cac_b6a1c4561ad6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a8966c_3784_4d97_9cac_b6a1c4561ad6.slice/crio-0952d34a8b0fe544ed8046ec0396eef638dde6d3d811210c3346511a0db492c4\": RecentStats: unable to find data in memory cache]" Oct 14 15:49:43 crc kubenswrapper[4945]: I1014 15:49:43.604324 4945 generic.go:334] "Generic (PLEG): container finished" podID="aa27100b-ae33-485e-b88d-3a695cab542a" containerID="5825161a9c57627e35d8abcb3e01cb81879973a91a04528dfa775ac2104428e4" exitCode=0 Oct 14 15:49:43 crc kubenswrapper[4945]: I1014 15:49:43.604469 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r9phx" event={"ID":"aa27100b-ae33-485e-b88d-3a695cab542a","Type":"ContainerDied","Data":"5825161a9c57627e35d8abcb3e01cb81879973a91a04528dfa775ac2104428e4"} Oct 14 15:49:44 crc kubenswrapper[4945]: I1014 15:49:44.615265 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r9phx" event={"ID":"aa27100b-ae33-485e-b88d-3a695cab542a","Type":"ContainerStarted","Data":"b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268"} Oct 14 15:49:44 crc kubenswrapper[4945]: I1014 15:49:44.642996 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r9phx" podStartSLOduration=3.048264748 podStartE2EDuration="4.64293342s" podCreationTimestamp="2025-10-14 15:49:40 +0000 UTC" firstStartedPulling="2025-10-14 15:49:42.585111816 +0000 UTC m=+2112.569160184" lastFinishedPulling="2025-10-14 15:49:44.179780488 +0000 UTC m=+2114.163828856" observedRunningTime="2025-10-14 15:49:44.633487921 +0000 UTC m=+2114.617536299" watchObservedRunningTime="2025-10-14 15:49:44.64293342 +0000 UTC m=+2114.626981788" Oct 14 15:49:45 crc kubenswrapper[4945]: I1014 15:49:45.267086 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7f6f97bcf4-qbb2j_9615a643-2f07-4637-b8d0-3acb4f24d0d8/barbican-api/0.log" Oct 14 15:49:45 crc kubenswrapper[4945]: I1014 15:49:45.434680 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7f6f97bcf4-qbb2j_9615a643-2f07-4637-b8d0-3acb4f24d0d8/barbican-api-log/0.log" Oct 14 15:49:45 crc kubenswrapper[4945]: I1014 15:49:45.453813 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-dfb99d5c8-zqfxx_e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67/barbican-keystone-listener/0.log" Oct 14 15:49:45 crc kubenswrapper[4945]: I1014 15:49:45.562562 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-dfb99d5c8-zqfxx_e7e2a3dd-f2e4-4b1c-8649-606fc87f5d67/barbican-keystone-listener-log/0.log" Oct 14 15:49:45 crc kubenswrapper[4945]: I1014 15:49:45.763980 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5749bb9b55-8z9wq_fcc68ba8-49a8-456d-b808-2d63dc0d6377/barbican-worker/0.log" Oct 14 15:49:45 crc kubenswrapper[4945]: I1014 15:49:45.885398 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5749bb9b55-8z9wq_fcc68ba8-49a8-456d-b808-2d63dc0d6377/barbican-worker-log/0.log" Oct 14 15:49:45 crc kubenswrapper[4945]: I1014 15:49:45.987389 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_519a62f5-6d77-4b41-9d34-52651d4e9571/ceilometer-central-agent/0.log" Oct 14 15:49:46 crc kubenswrapper[4945]: I1014 15:49:46.080175 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_519a62f5-6d77-4b41-9d34-52651d4e9571/proxy-httpd/0.log" Oct 14 15:49:46 crc kubenswrapper[4945]: I1014 15:49:46.097744 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_519a62f5-6d77-4b41-9d34-52651d4e9571/ceilometer-notification-agent/0.log" Oct 14 15:49:46 crc kubenswrapper[4945]: I1014 15:49:46.196096 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_519a62f5-6d77-4b41-9d34-52651d4e9571/sg-core/0.log" Oct 14 15:49:46 crc kubenswrapper[4945]: I1014 15:49:46.358484 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b210a0b1-468b-4e67-8618-02cb27c78033/cinder-api-log/0.log" Oct 14 15:49:46 crc kubenswrapper[4945]: I1014 15:49:46.364819 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b210a0b1-468b-4e67-8618-02cb27c78033/cinder-api/0.log" Oct 14 15:49:46 crc kubenswrapper[4945]: I1014 15:49:46.551680 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b15e8808-64a4-4e83-8a8a-1ff4ecc68711/cinder-scheduler/0.log" Oct 14 15:49:46 crc kubenswrapper[4945]: I1014 15:49:46.582364 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b15e8808-64a4-4e83-8a8a-1ff4ecc68711/probe/0.log" Oct 14 15:49:46 crc kubenswrapper[4945]: I1014 15:49:46.668995 4945 scope.go:117] "RemoveContainer" containerID="1e2872b7355c57f813dfa8bec061238ec7b3bce856ead60ed2ebe2fc124e69cc" Oct 14 15:49:46 crc kubenswrapper[4945]: I1014 15:49:46.874817 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5c7b6c5df9-hhtr4_9b5e61fe-2a22-4128-90c3-569e2e83a972/init/0.log" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.038955 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5c7b6c5df9-hhtr4_9b5e61fe-2a22-4128-90c3-569e2e83a972/init/0.log" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.083323 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5c7b6c5df9-hhtr4_9b5e61fe-2a22-4128-90c3-569e2e83a972/dnsmasq-dns/0.log" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.093581 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_57cbaab0-b5ce-4581-af58-87e3e3ec679c/glance-httpd/0.log" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.244501 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_57cbaab0-b5ce-4581-af58-87e3e3ec679c/glance-log/0.log" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.295761 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_314804da-3d6d-44cd-a9cf-0f014d690fe9/glance-httpd/0.log" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.339248 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_314804da-3d6d-44cd-a9cf-0f014d690fe9/glance-log/0.log" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.379444 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5lk5h"] Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.381883 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.398736 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5lk5h"] Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.435068 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6049943a-96c2-42d1-9da5-0fdebabafe3a-utilities\") pod \"redhat-operators-5lk5h\" (UID: \"6049943a-96c2-42d1-9da5-0fdebabafe3a\") " pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.435218 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf28z\" (UniqueName: \"kubernetes.io/projected/6049943a-96c2-42d1-9da5-0fdebabafe3a-kube-api-access-vf28z\") pod \"redhat-operators-5lk5h\" (UID: \"6049943a-96c2-42d1-9da5-0fdebabafe3a\") " pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.435389 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6049943a-96c2-42d1-9da5-0fdebabafe3a-catalog-content\") pod \"redhat-operators-5lk5h\" (UID: \"6049943a-96c2-42d1-9da5-0fdebabafe3a\") " pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.525214 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-bf87b4676-qpt54_5f9f45b7-0e91-4ecf-b38f-49bb07f40868/init/0.log" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.536627 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6049943a-96c2-42d1-9da5-0fdebabafe3a-utilities\") pod \"redhat-operators-5lk5h\" (UID: \"6049943a-96c2-42d1-9da5-0fdebabafe3a\") " pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.536754 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf28z\" (UniqueName: \"kubernetes.io/projected/6049943a-96c2-42d1-9da5-0fdebabafe3a-kube-api-access-vf28z\") pod \"redhat-operators-5lk5h\" (UID: \"6049943a-96c2-42d1-9da5-0fdebabafe3a\") " pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.536900 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6049943a-96c2-42d1-9da5-0fdebabafe3a-catalog-content\") pod \"redhat-operators-5lk5h\" (UID: \"6049943a-96c2-42d1-9da5-0fdebabafe3a\") " pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.537203 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6049943a-96c2-42d1-9da5-0fdebabafe3a-utilities\") pod \"redhat-operators-5lk5h\" (UID: \"6049943a-96c2-42d1-9da5-0fdebabafe3a\") " pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.537257 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6049943a-96c2-42d1-9da5-0fdebabafe3a-catalog-content\") pod \"redhat-operators-5lk5h\" (UID: \"6049943a-96c2-42d1-9da5-0fdebabafe3a\") " pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.559317 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf28z\" (UniqueName: \"kubernetes.io/projected/6049943a-96c2-42d1-9da5-0fdebabafe3a-kube-api-access-vf28z\") pod \"redhat-operators-5lk5h\" (UID: \"6049943a-96c2-42d1-9da5-0fdebabafe3a\") " pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.712042 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-bf87b4676-qpt54_5f9f45b7-0e91-4ecf-b38f-49bb07f40868/init/0.log" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.717129 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.725339 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-bf87b4676-qpt54_5f9f45b7-0e91-4ecf-b38f-49bb07f40868/ironic-api-log/0.log" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.793005 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-bf87b4676-qpt54_5f9f45b7-0e91-4ecf-b38f-49bb07f40868/ironic-api/0.log" Oct 14 15:49:47 crc kubenswrapper[4945]: I1014 15:49:47.984996 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/init/0.log" Oct 14 15:49:48 crc kubenswrapper[4945]: I1014 15:49:48.187707 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5lk5h"] Oct 14 15:49:48 crc kubenswrapper[4945]: W1014 15:49:48.192059 4945 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6049943a_96c2_42d1_9da5_0fdebabafe3a.slice/crio-eb847154351527bd1c1681cb5664a98b14f69bf5c30c41a0264e1753269b4a56 WatchSource:0}: Error finding container eb847154351527bd1c1681cb5664a98b14f69bf5c30c41a0264e1753269b4a56: Status 404 returned error can't find the container with id eb847154351527bd1c1681cb5664a98b14f69bf5c30c41a0264e1753269b4a56 Oct 14 15:49:48 crc kubenswrapper[4945]: I1014 15:49:48.327529 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/init/0.log" Oct 14 15:49:48 crc kubenswrapper[4945]: I1014 15:49:48.373499 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/ironic-python-agent-init/0.log" Oct 14 15:49:48 crc kubenswrapper[4945]: I1014 15:49:48.378059 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/ironic-python-agent-init/0.log" Oct 14 15:49:48 crc kubenswrapper[4945]: I1014 15:49:48.664157 4945 generic.go:334] "Generic (PLEG): container finished" podID="6049943a-96c2-42d1-9da5-0fdebabafe3a" containerID="1bbe2b2603e8c7e5583de86f4fbb0b034b345af13c2010b5534ca54b69257d0e" exitCode=0 Oct 14 15:49:48 crc kubenswrapper[4945]: I1014 15:49:48.664464 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lk5h" event={"ID":"6049943a-96c2-42d1-9da5-0fdebabafe3a","Type":"ContainerDied","Data":"1bbe2b2603e8c7e5583de86f4fbb0b034b345af13c2010b5534ca54b69257d0e"} Oct 14 15:49:48 crc kubenswrapper[4945]: I1014 15:49:48.664493 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lk5h" event={"ID":"6049943a-96c2-42d1-9da5-0fdebabafe3a","Type":"ContainerStarted","Data":"eb847154351527bd1c1681cb5664a98b14f69bf5c30c41a0264e1753269b4a56"} Oct 14 15:49:48 crc kubenswrapper[4945]: I1014 15:49:48.667787 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/init/0.log" Oct 14 15:49:48 crc kubenswrapper[4945]: I1014 15:49:48.749665 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/ironic-python-agent-init/0.log" Oct 14 15:49:49 crc kubenswrapper[4945]: I1014 15:49:49.037550 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/pxe-init/0.log" Oct 14 15:49:49 crc kubenswrapper[4945]: I1014 15:49:49.078204 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/init/0.log" Oct 14 15:49:49 crc kubenswrapper[4945]: I1014 15:49:49.249655 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/ironic-python-agent-init/0.log" Oct 14 15:49:49 crc kubenswrapper[4945]: I1014 15:49:49.332820 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/httpboot/0.log" Oct 14 15:49:49 crc kubenswrapper[4945]: I1014 15:49:49.535074 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/ironic-conductor/0.log" Oct 14 15:49:49 crc kubenswrapper[4945]: I1014 15:49:49.559647 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/ramdisk-logs/0.log" Oct 14 15:49:49 crc kubenswrapper[4945]: I1014 15:49:49.771620 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-gcqrl_367d71fe-5d2e-470a-958d-4a97ad15545c/init/0.log" Oct 14 15:49:49 crc kubenswrapper[4945]: I1014 15:49:49.976642 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-gcqrl_367d71fe-5d2e-470a-958d-4a97ad15545c/init/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.035794 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-gcqrl_367d71fe-5d2e-470a-958d-4a97ad15545c/ironic-db-sync/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.088176 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/pxe-init/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.224379 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/pxe-init/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.257945 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/ironic-python-agent-init/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.411080 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/ironic-python-agent-init/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.421292 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_042f4fc3-10ec-450e-b3cf-77a11d71a8b7/pxe-init/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.434677 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/inspector-pxe-init/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.476722 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/inspector-pxe-init/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.638817 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/inspector-httpboot/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.644907 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/ironic-python-agent-init/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.654772 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/inspector-pxe-init/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.677546 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/ironic-inspector/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.681762 4945 generic.go:334] "Generic (PLEG): container finished" podID="6049943a-96c2-42d1-9da5-0fdebabafe3a" containerID="1473a1ee5a2b364e304e16d83db5dda79a27af4adb574a29374ffb5bc1f0f3b9" exitCode=0 Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.681819 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lk5h" event={"ID":"6049943a-96c2-42d1-9da5-0fdebabafe3a","Type":"ContainerDied","Data":"1473a1ee5a2b364e304e16d83db5dda79a27af4adb574a29374ffb5bc1f0f3b9"} Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.727976 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/ironic-inspector-httpd/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.825635 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_3e3ee8d8-f91a-40ae-889c-b7e86a13f8f1/ramdisk-logs/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.851345 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-db-sync-t252w_048f7a68-56de-43d2-8ffb-33acacc4e2ec/ironic-inspector-db-sync/0.log" Oct 14 15:49:50 crc kubenswrapper[4945]: I1014 15:49:50.939768 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-neutron-agent-648bd876cc-srkfr_11506640-cea7-421c-878c-c4b449df880f/ironic-neutron-agent/3.log" Oct 14 15:49:51 crc kubenswrapper[4945]: I1014 15:49:51.009375 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-neutron-agent-648bd876cc-srkfr_11506640-cea7-421c-878c-c4b449df880f/ironic-neutron-agent/2.log" Oct 14 15:49:51 crc kubenswrapper[4945]: I1014 15:49:51.147811 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_41add807-ab1f-4d5f-999f-51626e4bb839/kube-state-metrics/0.log" Oct 14 15:49:51 crc kubenswrapper[4945]: I1014 15:49:51.221124 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5cdf875b4d-nbblw_fab41439-cbf7-47cd-a439-70519713d3db/keystone-api/0.log" Oct 14 15:49:51 crc kubenswrapper[4945]: I1014 15:49:51.330825 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:51 crc kubenswrapper[4945]: I1014 15:49:51.331094 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:51 crc kubenswrapper[4945]: I1014 15:49:51.385268 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:51 crc kubenswrapper[4945]: I1014 15:49:51.493319 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7cb8f5594f-b9dt5_904476f1-4b6d-432f-9db6-a0fd60f6ac3a/neutron-httpd/0.log" Oct 14 15:49:51 crc kubenswrapper[4945]: I1014 15:49:51.637054 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7cb8f5594f-b9dt5_904476f1-4b6d-432f-9db6-a0fd60f6ac3a/neutron-api/0.log" Oct 14 15:49:51 crc kubenswrapper[4945]: I1014 15:49:51.692675 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lk5h" event={"ID":"6049943a-96c2-42d1-9da5-0fdebabafe3a","Type":"ContainerStarted","Data":"ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2"} Oct 14 15:49:51 crc kubenswrapper[4945]: I1014 15:49:51.716640 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5lk5h" podStartSLOduration=2.212187449 podStartE2EDuration="4.716622734s" podCreationTimestamp="2025-10-14 15:49:47 +0000 UTC" firstStartedPulling="2025-10-14 15:49:48.668034 +0000 UTC m=+2118.652082368" lastFinishedPulling="2025-10-14 15:49:51.172469285 +0000 UTC m=+2121.156517653" observedRunningTime="2025-10-14 15:49:51.713841415 +0000 UTC m=+2121.697889803" watchObservedRunningTime="2025-10-14 15:49:51.716622734 +0000 UTC m=+2121.700671102" Oct 14 15:49:51 crc kubenswrapper[4945]: I1014 15:49:51.752177 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:51 crc kubenswrapper[4945]: I1014 15:49:51.926689 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a821c3a6-1324-4c0e-bcf4-1700cb3cfd69/nova-api-log/0.log" Oct 14 15:49:52 crc kubenswrapper[4945]: I1014 15:49:52.114413 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f4bf8e7b-6b0a-48da-831c-a9c307af75b9/nova-cell0-conductor-conductor/0.log" Oct 14 15:49:52 crc kubenswrapper[4945]: I1014 15:49:52.131807 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a821c3a6-1324-4c0e-bcf4-1700cb3cfd69/nova-api-api/0.log" Oct 14 15:49:52 crc kubenswrapper[4945]: I1014 15:49:52.322600 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_aaedddf9-b122-48e8-9dfc-e3525faaf9d3/nova-cell1-conductor-conductor/0.log" Oct 14 15:49:52 crc kubenswrapper[4945]: I1014 15:49:52.483831 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9af8e27a-69da-4094-925f-0d018b30d573/nova-cell1-novncproxy-novncproxy/0.log" Oct 14 15:49:52 crc kubenswrapper[4945]: I1014 15:49:52.547973 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f22b1bbe-8060-4302-b929-567dc4f33173/nova-metadata-log/0.log" Oct 14 15:49:52 crc kubenswrapper[4945]: I1014 15:49:52.909551 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f0fa2603-a8a2-449d-88d2-e35975a18ad1/mysql-bootstrap/0.log" Oct 14 15:49:52 crc kubenswrapper[4945]: I1014 15:49:52.914235 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_524fdd1e-9a42-495f-a692-e4722ab3c4e2/nova-scheduler-scheduler/0.log" Oct 14 15:49:53 crc kubenswrapper[4945]: I1014 15:49:53.062411 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f22b1bbe-8060-4302-b929-567dc4f33173/nova-metadata-metadata/0.log" Oct 14 15:49:53 crc kubenswrapper[4945]: I1014 15:49:53.165469 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f0fa2603-a8a2-449d-88d2-e35975a18ad1/galera/0.log" Oct 14 15:49:53 crc kubenswrapper[4945]: I1014 15:49:53.171954 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f0fa2603-a8a2-449d-88d2-e35975a18ad1/mysql-bootstrap/0.log" Oct 14 15:49:53 crc kubenswrapper[4945]: I1014 15:49:53.281962 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_afed9c20-1964-44dd-a849-af9fc0f97e86/mysql-bootstrap/0.log" Oct 14 15:49:53 crc kubenswrapper[4945]: I1014 15:49:53.377396 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r9phx"] Oct 14 15:49:53 crc kubenswrapper[4945]: E1014 15:49:53.434091 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a8966c_3784_4d97_9cac_b6a1c4561ad6.slice/crio-0952d34a8b0fe544ed8046ec0396eef638dde6d3d811210c3346511a0db492c4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a8966c_3784_4d97_9cac_b6a1c4561ad6.slice\": RecentStats: unable to find data in memory cache]" Oct 14 15:49:53 crc kubenswrapper[4945]: I1014 15:49:53.537716 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_473a83ad-e7a6-4b5b-899e-b64ef41465c9/openstackclient/0.log" Oct 14 15:49:53 crc kubenswrapper[4945]: I1014 15:49:53.543842 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_afed9c20-1964-44dd-a849-af9fc0f97e86/galera/0.log" Oct 14 15:49:53 crc kubenswrapper[4945]: I1014 15:49:53.575237 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_afed9c20-1964-44dd-a849-af9fc0f97e86/mysql-bootstrap/0.log" Oct 14 15:49:53 crc kubenswrapper[4945]: I1014 15:49:53.712051 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r9phx" podUID="aa27100b-ae33-485e-b88d-3a695cab542a" containerName="registry-server" containerID="cri-o://b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268" gracePeriod=2 Oct 14 15:49:53 crc kubenswrapper[4945]: I1014 15:49:53.843248 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-krh6j_80e8595d-2cf1-4fcf-89c1-feed0e8ef1f4/openstack-network-exporter/0.log" Oct 14 15:49:53 crc kubenswrapper[4945]: I1014 15:49:53.903541 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6v9cw_9d0edbee-18bd-41ac-880b-526477ef54d3/ovn-controller/0.log" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.080832 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pq8xf_2f495edd-59f6-4a56-9356-496cda4ae47a/ovsdb-server-init/0.log" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.217258 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.321082 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pq8xf_2f495edd-59f6-4a56-9356-496cda4ae47a/ovsdb-server-init/0.log" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.362645 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9254\" (UniqueName: \"kubernetes.io/projected/aa27100b-ae33-485e-b88d-3a695cab542a-kube-api-access-z9254\") pod \"aa27100b-ae33-485e-b88d-3a695cab542a\" (UID: \"aa27100b-ae33-485e-b88d-3a695cab542a\") " Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.362751 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa27100b-ae33-485e-b88d-3a695cab542a-catalog-content\") pod \"aa27100b-ae33-485e-b88d-3a695cab542a\" (UID: \"aa27100b-ae33-485e-b88d-3a695cab542a\") " Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.362916 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa27100b-ae33-485e-b88d-3a695cab542a-utilities\") pod \"aa27100b-ae33-485e-b88d-3a695cab542a\" (UID: \"aa27100b-ae33-485e-b88d-3a695cab542a\") " Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.366219 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa27100b-ae33-485e-b88d-3a695cab542a-utilities" (OuterVolumeSpecName: "utilities") pod "aa27100b-ae33-485e-b88d-3a695cab542a" (UID: "aa27100b-ae33-485e-b88d-3a695cab542a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.390158 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa27100b-ae33-485e-b88d-3a695cab542a-kube-api-access-z9254" (OuterVolumeSpecName: "kube-api-access-z9254") pod "aa27100b-ae33-485e-b88d-3a695cab542a" (UID: "aa27100b-ae33-485e-b88d-3a695cab542a"). InnerVolumeSpecName "kube-api-access-z9254". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.393503 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa27100b-ae33-485e-b88d-3a695cab542a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa27100b-ae33-485e-b88d-3a695cab542a" (UID: "aa27100b-ae33-485e-b88d-3a695cab542a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.400680 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pq8xf_2f495edd-59f6-4a56-9356-496cda4ae47a/ovsdb-server/0.log" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.429465 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pq8xf_2f495edd-59f6-4a56-9356-496cda4ae47a/ovs-vswitchd/0.log" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.465336 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa27100b-ae33-485e-b88d-3a695cab542a-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.465364 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9254\" (UniqueName: \"kubernetes.io/projected/aa27100b-ae33-485e-b88d-3a695cab542a-kube-api-access-z9254\") on node \"crc\" DevicePath \"\"" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.465375 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa27100b-ae33-485e-b88d-3a695cab542a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.626418 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31/openstack-network-exporter/0.log" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.644603 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3d7482e7-31e7-4d3f-ae1c-84b8ce31ed31/ovn-northd/0.log" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.738594 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e90e1239-2529-43dd-b429-c140543cb579/openstack-network-exporter/0.log" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.739488 4945 generic.go:334] "Generic (PLEG): container finished" podID="aa27100b-ae33-485e-b88d-3a695cab542a" containerID="b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268" exitCode=0 Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.739534 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r9phx" event={"ID":"aa27100b-ae33-485e-b88d-3a695cab542a","Type":"ContainerDied","Data":"b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268"} Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.739564 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r9phx" event={"ID":"aa27100b-ae33-485e-b88d-3a695cab542a","Type":"ContainerDied","Data":"4b783fcc116091cae372f5f971a202bb06c6248a150bad93a427ac9998e04acc"} Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.739581 4945 scope.go:117] "RemoveContainer" containerID="b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.739707 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r9phx" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.770528 4945 scope.go:117] "RemoveContainer" containerID="5825161a9c57627e35d8abcb3e01cb81879973a91a04528dfa775ac2104428e4" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.782344 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r9phx"] Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.787390 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r9phx"] Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.790920 4945 scope.go:117] "RemoveContainer" containerID="1e89d26ce41df9312296ac9fd8a26c61f4a2c828d9aefd4d2f36f660945906e4" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.869485 4945 scope.go:117] "RemoveContainer" containerID="b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268" Oct 14 15:49:54 crc kubenswrapper[4945]: E1014 15:49:54.872968 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268\": container with ID starting with b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268 not found: ID does not exist" containerID="b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.873106 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268"} err="failed to get container status \"b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268\": rpc error: code = NotFound desc = could not find container \"b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268\": container with ID starting with b1948510521c43e3455cc515fcae994ea3d51a75ae34292a0f183f20d9035268 not found: ID does not exist" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.873221 4945 scope.go:117] "RemoveContainer" containerID="5825161a9c57627e35d8abcb3e01cb81879973a91a04528dfa775ac2104428e4" Oct 14 15:49:54 crc kubenswrapper[4945]: E1014 15:49:54.877007 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5825161a9c57627e35d8abcb3e01cb81879973a91a04528dfa775ac2104428e4\": container with ID starting with 5825161a9c57627e35d8abcb3e01cb81879973a91a04528dfa775ac2104428e4 not found: ID does not exist" containerID="5825161a9c57627e35d8abcb3e01cb81879973a91a04528dfa775ac2104428e4" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.877045 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5825161a9c57627e35d8abcb3e01cb81879973a91a04528dfa775ac2104428e4"} err="failed to get container status \"5825161a9c57627e35d8abcb3e01cb81879973a91a04528dfa775ac2104428e4\": rpc error: code = NotFound desc = could not find container \"5825161a9c57627e35d8abcb3e01cb81879973a91a04528dfa775ac2104428e4\": container with ID starting with 5825161a9c57627e35d8abcb3e01cb81879973a91a04528dfa775ac2104428e4 not found: ID does not exist" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.877110 4945 scope.go:117] "RemoveContainer" containerID="1e89d26ce41df9312296ac9fd8a26c61f4a2c828d9aefd4d2f36f660945906e4" Oct 14 15:49:54 crc kubenswrapper[4945]: E1014 15:49:54.877488 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e89d26ce41df9312296ac9fd8a26c61f4a2c828d9aefd4d2f36f660945906e4\": container with ID starting with 1e89d26ce41df9312296ac9fd8a26c61f4a2c828d9aefd4d2f36f660945906e4 not found: ID does not exist" containerID="1e89d26ce41df9312296ac9fd8a26c61f4a2c828d9aefd4d2f36f660945906e4" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.877518 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e89d26ce41df9312296ac9fd8a26c61f4a2c828d9aefd4d2f36f660945906e4"} err="failed to get container status \"1e89d26ce41df9312296ac9fd8a26c61f4a2c828d9aefd4d2f36f660945906e4\": rpc error: code = NotFound desc = could not find container \"1e89d26ce41df9312296ac9fd8a26c61f4a2c828d9aefd4d2f36f660945906e4\": container with ID starting with 1e89d26ce41df9312296ac9fd8a26c61f4a2c828d9aefd4d2f36f660945906e4 not found: ID does not exist" Oct 14 15:49:54 crc kubenswrapper[4945]: I1014 15:49:54.902204 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e90e1239-2529-43dd-b429-c140543cb579/ovsdbserver-nb/0.log" Oct 14 15:49:55 crc kubenswrapper[4945]: I1014 15:49:55.007263 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3/openstack-network-exporter/0.log" Oct 14 15:49:55 crc kubenswrapper[4945]: I1014 15:49:55.030845 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2195ac97-ef9f-4fd8-b56b-4b96bb1c59f3/ovsdbserver-sb/0.log" Oct 14 15:49:55 crc kubenswrapper[4945]: I1014 15:49:55.276011 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-bd8df944d-q4z6c_396db799-75e0-4156-a97d-472db3c42729/placement-api/0.log" Oct 14 15:49:55 crc kubenswrapper[4945]: I1014 15:49:55.430371 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-bd8df944d-q4z6c_396db799-75e0-4156-a97d-472db3c42729/placement-log/0.log" Oct 14 15:49:55 crc kubenswrapper[4945]: I1014 15:49:55.463822 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b52be853-0c59-48be-b4d0-3b2023310050/setup-container/0.log" Oct 14 15:49:55 crc kubenswrapper[4945]: I1014 15:49:55.774193 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b52be853-0c59-48be-b4d0-3b2023310050/setup-container/0.log" Oct 14 15:49:55 crc kubenswrapper[4945]: I1014 15:49:55.844829 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b52be853-0c59-48be-b4d0-3b2023310050/rabbitmq/0.log" Oct 14 15:49:55 crc kubenswrapper[4945]: I1014 15:49:55.848758 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_40cdc335-20ee-467f-bc15-4c5867d0fd2a/setup-container/0.log" Oct 14 15:49:55 crc kubenswrapper[4945]: I1014 15:49:55.990404 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_40cdc335-20ee-467f-bc15-4c5867d0fd2a/setup-container/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.125583 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_40cdc335-20ee-467f-bc15-4c5867d0fd2a/rabbitmq/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.211330 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-86b568845c-dgcjh_047520e2-5898-42bd-85bf-dce4063f08f4/proxy-httpd/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.312271 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-86b568845c-dgcjh_047520e2-5898-42bd-85bf-dce4063f08f4/proxy-server/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.403258 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-h5b5w_787ea529-5875-4e6c-8a3e-5b4fbf019fa8/swift-ring-rebalance/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.519033 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/account-auditor/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.564108 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/account-reaper/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.624084 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/account-replicator/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.682932 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/account-server/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.765836 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/container-auditor/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.773813 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa27100b-ae33-485e-b88d-3a695cab542a" path="/var/lib/kubelet/pods/aa27100b-ae33-485e-b88d-3a695cab542a/volumes" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.776856 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/container-replicator/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.848867 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/container-server/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.886355 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/container-updater/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.987440 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/object-auditor/0.log" Oct 14 15:49:56 crc kubenswrapper[4945]: I1014 15:49:56.997642 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/object-expirer/0.log" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.104805 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/object-replicator/0.log" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.117518 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/object-server/0.log" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.224943 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/object-updater/0.log" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.275780 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/rsync/0.log" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.379935 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_eb8ab92b-e774-40b3-becd-458958582bef/swift-recon-cron/0.log" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.717231 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.717921 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.787091 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8kt6g"] Oct 14 15:49:57 crc kubenswrapper[4945]: E1014 15:49:57.787493 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa27100b-ae33-485e-b88d-3a695cab542a" containerName="extract-content" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.787508 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa27100b-ae33-485e-b88d-3a695cab542a" containerName="extract-content" Oct 14 15:49:57 crc kubenswrapper[4945]: E1014 15:49:57.787546 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa27100b-ae33-485e-b88d-3a695cab542a" containerName="registry-server" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.787552 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa27100b-ae33-485e-b88d-3a695cab542a" containerName="registry-server" Oct 14 15:49:57 crc kubenswrapper[4945]: E1014 15:49:57.787570 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa27100b-ae33-485e-b88d-3a695cab542a" containerName="extract-utilities" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.787577 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa27100b-ae33-485e-b88d-3a695cab542a" containerName="extract-utilities" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.787754 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa27100b-ae33-485e-b88d-3a695cab542a" containerName="registry-server" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.793919 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.794856 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.816411 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8kt6g"] Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.941468 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9de0e862-25dd-46e1-9173-2419826478a8-utilities\") pod \"certified-operators-8kt6g\" (UID: \"9de0e862-25dd-46e1-9173-2419826478a8\") " pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.941512 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tjwj\" (UniqueName: \"kubernetes.io/projected/9de0e862-25dd-46e1-9173-2419826478a8-kube-api-access-4tjwj\") pod \"certified-operators-8kt6g\" (UID: \"9de0e862-25dd-46e1-9173-2419826478a8\") " pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:49:57 crc kubenswrapper[4945]: I1014 15:49:57.941572 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9de0e862-25dd-46e1-9173-2419826478a8-catalog-content\") pod \"certified-operators-8kt6g\" (UID: \"9de0e862-25dd-46e1-9173-2419826478a8\") " pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:49:58 crc kubenswrapper[4945]: I1014 15:49:58.043076 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9de0e862-25dd-46e1-9173-2419826478a8-utilities\") pod \"certified-operators-8kt6g\" (UID: \"9de0e862-25dd-46e1-9173-2419826478a8\") " pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:49:58 crc kubenswrapper[4945]: I1014 15:49:58.043436 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tjwj\" (UniqueName: \"kubernetes.io/projected/9de0e862-25dd-46e1-9173-2419826478a8-kube-api-access-4tjwj\") pod \"certified-operators-8kt6g\" (UID: \"9de0e862-25dd-46e1-9173-2419826478a8\") " pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:49:58 crc kubenswrapper[4945]: I1014 15:49:58.043507 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9de0e862-25dd-46e1-9173-2419826478a8-catalog-content\") pod \"certified-operators-8kt6g\" (UID: \"9de0e862-25dd-46e1-9173-2419826478a8\") " pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:49:58 crc kubenswrapper[4945]: I1014 15:49:58.043514 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9de0e862-25dd-46e1-9173-2419826478a8-utilities\") pod \"certified-operators-8kt6g\" (UID: \"9de0e862-25dd-46e1-9173-2419826478a8\") " pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:49:58 crc kubenswrapper[4945]: I1014 15:49:58.043844 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9de0e862-25dd-46e1-9173-2419826478a8-catalog-content\") pod \"certified-operators-8kt6g\" (UID: \"9de0e862-25dd-46e1-9173-2419826478a8\") " pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:49:58 crc kubenswrapper[4945]: I1014 15:49:58.065542 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tjwj\" (UniqueName: \"kubernetes.io/projected/9de0e862-25dd-46e1-9173-2419826478a8-kube-api-access-4tjwj\") pod \"certified-operators-8kt6g\" (UID: \"9de0e862-25dd-46e1-9173-2419826478a8\") " pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:49:58 crc kubenswrapper[4945]: I1014 15:49:58.126897 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:49:58 crc kubenswrapper[4945]: I1014 15:49:58.730588 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8kt6g"] Oct 14 15:49:58 crc kubenswrapper[4945]: I1014 15:49:58.848016 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kt6g" event={"ID":"9de0e862-25dd-46e1-9173-2419826478a8","Type":"ContainerStarted","Data":"c23afa0c398e7b99b6538112fb0f10dcbd71e434d8d37150e14d1e5ec7dc7780"} Oct 14 15:49:58 crc kubenswrapper[4945]: I1014 15:49:58.908288 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:49:59 crc kubenswrapper[4945]: I1014 15:49:59.878336 4945 generic.go:334] "Generic (PLEG): container finished" podID="9de0e862-25dd-46e1-9173-2419826478a8" containerID="2cdba31bc8248fc5e104c7e91945488cbf946da29e91eeafa4e8888aa20f46b2" exitCode=0 Oct 14 15:49:59 crc kubenswrapper[4945]: I1014 15:49:59.878460 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kt6g" event={"ID":"9de0e862-25dd-46e1-9173-2419826478a8","Type":"ContainerDied","Data":"2cdba31bc8248fc5e104c7e91945488cbf946da29e91eeafa4e8888aa20f46b2"} Oct 14 15:50:00 crc kubenswrapper[4945]: I1014 15:50:00.172864 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5lk5h"] Oct 14 15:50:00 crc kubenswrapper[4945]: I1014 15:50:00.885646 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5lk5h" podUID="6049943a-96c2-42d1-9da5-0fdebabafe3a" containerName="registry-server" containerID="cri-o://ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2" gracePeriod=2 Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.552931 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.640442 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6049943a-96c2-42d1-9da5-0fdebabafe3a-utilities\") pod \"6049943a-96c2-42d1-9da5-0fdebabafe3a\" (UID: \"6049943a-96c2-42d1-9da5-0fdebabafe3a\") " Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.640658 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6049943a-96c2-42d1-9da5-0fdebabafe3a-catalog-content\") pod \"6049943a-96c2-42d1-9da5-0fdebabafe3a\" (UID: \"6049943a-96c2-42d1-9da5-0fdebabafe3a\") " Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.640755 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf28z\" (UniqueName: \"kubernetes.io/projected/6049943a-96c2-42d1-9da5-0fdebabafe3a-kube-api-access-vf28z\") pod \"6049943a-96c2-42d1-9da5-0fdebabafe3a\" (UID: \"6049943a-96c2-42d1-9da5-0fdebabafe3a\") " Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.647020 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6049943a-96c2-42d1-9da5-0fdebabafe3a-kube-api-access-vf28z" (OuterVolumeSpecName: "kube-api-access-vf28z") pod "6049943a-96c2-42d1-9da5-0fdebabafe3a" (UID: "6049943a-96c2-42d1-9da5-0fdebabafe3a"). InnerVolumeSpecName "kube-api-access-vf28z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.652791 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6049943a-96c2-42d1-9da5-0fdebabafe3a-utilities" (OuterVolumeSpecName: "utilities") pod "6049943a-96c2-42d1-9da5-0fdebabafe3a" (UID: "6049943a-96c2-42d1-9da5-0fdebabafe3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.742555 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf28z\" (UniqueName: \"kubernetes.io/projected/6049943a-96c2-42d1-9da5-0fdebabafe3a-kube-api-access-vf28z\") on node \"crc\" DevicePath \"\"" Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.742755 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6049943a-96c2-42d1-9da5-0fdebabafe3a-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.772116 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6049943a-96c2-42d1-9da5-0fdebabafe3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6049943a-96c2-42d1-9da5-0fdebabafe3a" (UID: "6049943a-96c2-42d1-9da5-0fdebabafe3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.844831 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6049943a-96c2-42d1-9da5-0fdebabafe3a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.922720 4945 generic.go:334] "Generic (PLEG): container finished" podID="6049943a-96c2-42d1-9da5-0fdebabafe3a" containerID="ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2" exitCode=0 Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.922766 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lk5h" event={"ID":"6049943a-96c2-42d1-9da5-0fdebabafe3a","Type":"ContainerDied","Data":"ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2"} Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.922795 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5lk5h" event={"ID":"6049943a-96c2-42d1-9da5-0fdebabafe3a","Type":"ContainerDied","Data":"eb847154351527bd1c1681cb5664a98b14f69bf5c30c41a0264e1753269b4a56"} Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.922814 4945 scope.go:117] "RemoveContainer" containerID="ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2" Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.922988 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5lk5h" Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.959038 4945 scope.go:117] "RemoveContainer" containerID="1473a1ee5a2b364e304e16d83db5dda79a27af4adb574a29374ffb5bc1f0f3b9" Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.958972 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5lk5h"] Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.967638 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5lk5h"] Oct 14 15:50:02 crc kubenswrapper[4945]: I1014 15:50:02.999450 4945 scope.go:117] "RemoveContainer" containerID="1bbe2b2603e8c7e5583de86f4fbb0b034b345af13c2010b5534ca54b69257d0e" Oct 14 15:50:03 crc kubenswrapper[4945]: I1014 15:50:03.049149 4945 scope.go:117] "RemoveContainer" containerID="ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2" Oct 14 15:50:03 crc kubenswrapper[4945]: E1014 15:50:03.052298 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2\": container with ID starting with ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2 not found: ID does not exist" containerID="ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2" Oct 14 15:50:03 crc kubenswrapper[4945]: I1014 15:50:03.052341 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2"} err="failed to get container status \"ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2\": rpc error: code = NotFound desc = could not find container \"ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2\": container with ID starting with ef0e5e1920250064007e21630c80b867a2c6baae2bf446d909852143c47b7dd2 not found: ID does not exist" Oct 14 15:50:03 crc kubenswrapper[4945]: I1014 15:50:03.052367 4945 scope.go:117] "RemoveContainer" containerID="1473a1ee5a2b364e304e16d83db5dda79a27af4adb574a29374ffb5bc1f0f3b9" Oct 14 15:50:03 crc kubenswrapper[4945]: E1014 15:50:03.052863 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1473a1ee5a2b364e304e16d83db5dda79a27af4adb574a29374ffb5bc1f0f3b9\": container with ID starting with 1473a1ee5a2b364e304e16d83db5dda79a27af4adb574a29374ffb5bc1f0f3b9 not found: ID does not exist" containerID="1473a1ee5a2b364e304e16d83db5dda79a27af4adb574a29374ffb5bc1f0f3b9" Oct 14 15:50:03 crc kubenswrapper[4945]: I1014 15:50:03.052900 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1473a1ee5a2b364e304e16d83db5dda79a27af4adb574a29374ffb5bc1f0f3b9"} err="failed to get container status \"1473a1ee5a2b364e304e16d83db5dda79a27af4adb574a29374ffb5bc1f0f3b9\": rpc error: code = NotFound desc = could not find container \"1473a1ee5a2b364e304e16d83db5dda79a27af4adb574a29374ffb5bc1f0f3b9\": container with ID starting with 1473a1ee5a2b364e304e16d83db5dda79a27af4adb574a29374ffb5bc1f0f3b9 not found: ID does not exist" Oct 14 15:50:03 crc kubenswrapper[4945]: I1014 15:50:03.052922 4945 scope.go:117] "RemoveContainer" containerID="1bbe2b2603e8c7e5583de86f4fbb0b034b345af13c2010b5534ca54b69257d0e" Oct 14 15:50:03 crc kubenswrapper[4945]: E1014 15:50:03.053547 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bbe2b2603e8c7e5583de86f4fbb0b034b345af13c2010b5534ca54b69257d0e\": container with ID starting with 1bbe2b2603e8c7e5583de86f4fbb0b034b345af13c2010b5534ca54b69257d0e not found: ID does not exist" containerID="1bbe2b2603e8c7e5583de86f4fbb0b034b345af13c2010b5534ca54b69257d0e" Oct 14 15:50:03 crc kubenswrapper[4945]: I1014 15:50:03.053591 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bbe2b2603e8c7e5583de86f4fbb0b034b345af13c2010b5534ca54b69257d0e"} err="failed to get container status \"1bbe2b2603e8c7e5583de86f4fbb0b034b345af13c2010b5534ca54b69257d0e\": rpc error: code = NotFound desc = could not find container \"1bbe2b2603e8c7e5583de86f4fbb0b034b345af13c2010b5534ca54b69257d0e\": container with ID starting with 1bbe2b2603e8c7e5583de86f4fbb0b034b345af13c2010b5534ca54b69257d0e not found: ID does not exist" Oct 14 15:50:03 crc kubenswrapper[4945]: I1014 15:50:03.267209 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_131ae7de-5f52-4fd7-8789-d61f8a3a9c46/memcached/0.log" Oct 14 15:50:03 crc kubenswrapper[4945]: E1014 15:50:03.773743 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a8966c_3784_4d97_9cac_b6a1c4561ad6.slice/crio-0952d34a8b0fe544ed8046ec0396eef638dde6d3d811210c3346511a0db492c4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a8966c_3784_4d97_9cac_b6a1c4561ad6.slice\": RecentStats: unable to find data in memory cache]" Oct 14 15:50:04 crc kubenswrapper[4945]: I1014 15:50:04.774678 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6049943a-96c2-42d1-9da5-0fdebabafe3a" path="/var/lib/kubelet/pods/6049943a-96c2-42d1-9da5-0fdebabafe3a/volumes" Oct 14 15:50:08 crc kubenswrapper[4945]: I1014 15:50:08.987036 4945 generic.go:334] "Generic (PLEG): container finished" podID="9de0e862-25dd-46e1-9173-2419826478a8" containerID="e7696cee03be80d3d166d3353fde870a4c363df0a4a3e0ddf719bd3049b78341" exitCode=0 Oct 14 15:50:08 crc kubenswrapper[4945]: I1014 15:50:08.987110 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kt6g" event={"ID":"9de0e862-25dd-46e1-9173-2419826478a8","Type":"ContainerDied","Data":"e7696cee03be80d3d166d3353fde870a4c363df0a4a3e0ddf719bd3049b78341"} Oct 14 15:50:09 crc kubenswrapper[4945]: I1014 15:50:09.996853 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kt6g" event={"ID":"9de0e862-25dd-46e1-9173-2419826478a8","Type":"ContainerStarted","Data":"cca470af5952a663233417cc2351d734d675610675d8b23beab503d5e51f8378"} Oct 14 15:50:10 crc kubenswrapper[4945]: I1014 15:50:10.025035 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8kt6g" podStartSLOduration=3.201372752 podStartE2EDuration="13.025015964s" podCreationTimestamp="2025-10-14 15:49:57 +0000 UTC" firstStartedPulling="2025-10-14 15:49:59.880401749 +0000 UTC m=+2129.864450117" lastFinishedPulling="2025-10-14 15:50:09.704044971 +0000 UTC m=+2139.688093329" observedRunningTime="2025-10-14 15:50:10.017739236 +0000 UTC m=+2140.001787624" watchObservedRunningTime="2025-10-14 15:50:10.025015964 +0000 UTC m=+2140.009064332" Oct 14 15:50:14 crc kubenswrapper[4945]: E1014 15:50:14.036902 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a8966c_3784_4d97_9cac_b6a1c4561ad6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a8966c_3784_4d97_9cac_b6a1c4561ad6.slice/crio-0952d34a8b0fe544ed8046ec0396eef638dde6d3d811210c3346511a0db492c4\": RecentStats: unable to find data in memory cache]" Oct 14 15:50:16 crc kubenswrapper[4945]: I1014 15:50:16.795378 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:50:16 crc kubenswrapper[4945]: I1014 15:50:16.795671 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:50:18 crc kubenswrapper[4945]: I1014 15:50:18.127861 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:50:18 crc kubenswrapper[4945]: I1014 15:50:18.128184 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:50:18 crc kubenswrapper[4945]: I1014 15:50:18.181297 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:50:19 crc kubenswrapper[4945]: I1014 15:50:19.150239 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8kt6g" Oct 14 15:50:19 crc kubenswrapper[4945]: I1014 15:50:19.231102 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8kt6g"] Oct 14 15:50:19 crc kubenswrapper[4945]: I1014 15:50:19.277691 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n87wl"] Oct 14 15:50:19 crc kubenswrapper[4945]: I1014 15:50:19.277965 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n87wl" podUID="b3811f6a-94ba-4d10-bd4b-873af175eec4" containerName="registry-server" containerID="cri-o://ed8b248b128314b208cd5836cc79025ce58f554e100fc87e8bfa6dbb75d22042" gracePeriod=2 Oct 14 15:50:19 crc kubenswrapper[4945]: I1014 15:50:19.695816 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-xlwn6_f5ac9c23-4d14-460d-9255-16d1865e9aa1/manager/0.log" Oct 14 15:50:19 crc kubenswrapper[4945]: I1014 15:50:19.722633 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-xlwn6_f5ac9c23-4d14-460d-9255-16d1865e9aa1/kube-rbac-proxy/0.log" Oct 14 15:50:19 crc kubenswrapper[4945]: I1014 15:50:19.972385 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-8gm4x_3ae6c6f4-2d25-420b-8149-a976de545cd6/kube-rbac-proxy/0.log" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.108294 4945 generic.go:334] "Generic (PLEG): container finished" podID="b3811f6a-94ba-4d10-bd4b-873af175eec4" containerID="ed8b248b128314b208cd5836cc79025ce58f554e100fc87e8bfa6dbb75d22042" exitCode=0 Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.108864 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n87wl" event={"ID":"b3811f6a-94ba-4d10-bd4b-873af175eec4","Type":"ContainerDied","Data":"ed8b248b128314b208cd5836cc79025ce58f554e100fc87e8bfa6dbb75d22042"} Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.134611 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-8gm4x_3ae6c6f4-2d25-420b-8149-a976de545cd6/manager/0.log" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.291257 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-gf66t_64596f58-bc5a-4cdf-80d4-22002c0c9575/manager/0.log" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.339289 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.415196 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-gf66t_64596f58-bc5a-4cdf-80d4-22002c0c9575/kube-rbac-proxy/0.log" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.507773 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6gpz\" (UniqueName: \"kubernetes.io/projected/b3811f6a-94ba-4d10-bd4b-873af175eec4-kube-api-access-j6gpz\") pod \"b3811f6a-94ba-4d10-bd4b-873af175eec4\" (UID: \"b3811f6a-94ba-4d10-bd4b-873af175eec4\") " Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.507831 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3811f6a-94ba-4d10-bd4b-873af175eec4-utilities\") pod \"b3811f6a-94ba-4d10-bd4b-873af175eec4\" (UID: \"b3811f6a-94ba-4d10-bd4b-873af175eec4\") " Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.507909 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3811f6a-94ba-4d10-bd4b-873af175eec4-catalog-content\") pod \"b3811f6a-94ba-4d10-bd4b-873af175eec4\" (UID: \"b3811f6a-94ba-4d10-bd4b-873af175eec4\") " Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.508497 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3811f6a-94ba-4d10-bd4b-873af175eec4-utilities" (OuterVolumeSpecName: "utilities") pod "b3811f6a-94ba-4d10-bd4b-873af175eec4" (UID: "b3811f6a-94ba-4d10-bd4b-873af175eec4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.512632 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3811f6a-94ba-4d10-bd4b-873af175eec4-kube-api-access-j6gpz" (OuterVolumeSpecName: "kube-api-access-j6gpz") pod "b3811f6a-94ba-4d10-bd4b-873af175eec4" (UID: "b3811f6a-94ba-4d10-bd4b-873af175eec4"). InnerVolumeSpecName "kube-api-access-j6gpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.513319 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6gpz\" (UniqueName: \"kubernetes.io/projected/b3811f6a-94ba-4d10-bd4b-873af175eec4-kube-api-access-j6gpz\") on node \"crc\" DevicePath \"\"" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.513349 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3811f6a-94ba-4d10-bd4b-873af175eec4-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.542566 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/util/0.log" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.579982 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3811f6a-94ba-4d10-bd4b-873af175eec4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3811f6a-94ba-4d10-bd4b-873af175eec4" (UID: "b3811f6a-94ba-4d10-bd4b-873af175eec4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.615770 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3811f6a-94ba-4d10-bd4b-873af175eec4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.823787 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/pull/0.log" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.826863 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/pull/0.log" Oct 14 15:50:20 crc kubenswrapper[4945]: I1014 15:50:20.863131 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/util/0.log" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.008363 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/util/0.log" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.014255 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/pull/0.log" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.046309 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e0420e9c71d23bd1575471989ef85cb83902801f70006223dfa97ca779ln8md_d84b1643-7a42-4df5-ad42-d31cf321ce9d/extract/0.log" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.121455 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n87wl" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.121454 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n87wl" event={"ID":"b3811f6a-94ba-4d10-bd4b-873af175eec4","Type":"ContainerDied","Data":"91109345b8bf767889700421d5e17426ba557827b65dd669074795fed002d656"} Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.121542 4945 scope.go:117] "RemoveContainer" containerID="ed8b248b128314b208cd5836cc79025ce58f554e100fc87e8bfa6dbb75d22042" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.142260 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n87wl"] Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.149359 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n87wl"] Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.150322 4945 scope.go:117] "RemoveContainer" containerID="2a247e4a6f832c55d0e1f9d0a4bf8115fe086e8a8a1fd149a9e856d8ab53b622" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.173735 4945 scope.go:117] "RemoveContainer" containerID="a7525fbf5d89d0f12d3ce93c0c869d45f9eb0c5a4aa424c894fcedb7769460ae" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.225198 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-6fg2j_86d698df-b089-498a-98e5-65c196aabcc3/kube-rbac-proxy/0.log" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.290326 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-v9hhr_0a17b021-e8aa-4f88-b150-596885e48ab1/kube-rbac-proxy/0.log" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.297487 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-6fg2j_86d698df-b089-498a-98e5-65c196aabcc3/manager/0.log" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.452672 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-v9hhr_0a17b021-e8aa-4f88-b150-596885e48ab1/manager/0.log" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.513534 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-wchqt_80c510fc-48a8-44c1-9927-c9b0ce934410/kube-rbac-proxy/0.log" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.569172 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-wchqt_80c510fc-48a8-44c1-9927-c9b0ce934410/manager/0.log" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.686484 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-pk25j_41d187ec-0816-4800-9d76-9d7a3f364391/kube-rbac-proxy/0.log" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.910607 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-76ccf76b65-p9trl_20fbc887-9299-460d-9890-f9d7e0235118/kube-rbac-proxy/0.log" Oct 14 15:50:21 crc kubenswrapper[4945]: I1014 15:50:21.946576 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-pk25j_41d187ec-0816-4800-9d76-9d7a3f364391/manager/0.log" Oct 14 15:50:22 crc kubenswrapper[4945]: I1014 15:50:22.000231 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-76ccf76b65-p9trl_20fbc887-9299-460d-9890-f9d7e0235118/manager/0.log" Oct 14 15:50:22 crc kubenswrapper[4945]: I1014 15:50:22.330838 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-fqb29_97138e7c-7207-4c47-898f-f71625e0e49b/kube-rbac-proxy/0.log" Oct 14 15:50:22 crc kubenswrapper[4945]: I1014 15:50:22.376160 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-fqb29_97138e7c-7207-4c47-898f-f71625e0e49b/manager/0.log" Oct 14 15:50:22 crc kubenswrapper[4945]: I1014 15:50:22.473504 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-jcrs7_2a652464-b536-4402-a883-f52390e0327b/kube-rbac-proxy/0.log" Oct 14 15:50:22 crc kubenswrapper[4945]: I1014 15:50:22.508375 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-jcrs7_2a652464-b536-4402-a883-f52390e0327b/manager/0.log" Oct 14 15:50:22 crc kubenswrapper[4945]: I1014 15:50:22.615430 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-h2p64_d39181a7-852e-402c-9912-7c81a0952e90/kube-rbac-proxy/0.log" Oct 14 15:50:22 crc kubenswrapper[4945]: I1014 15:50:22.732863 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-h2p64_d39181a7-852e-402c-9912-7c81a0952e90/manager/0.log" Oct 14 15:50:22 crc kubenswrapper[4945]: I1014 15:50:22.750054 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-qx6hl_d10e063a-f99e-4fc8-b539-224bbf23f466/kube-rbac-proxy/0.log" Oct 14 15:50:22 crc kubenswrapper[4945]: I1014 15:50:22.775396 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3811f6a-94ba-4d10-bd4b-873af175eec4" path="/var/lib/kubelet/pods/b3811f6a-94ba-4d10-bd4b-873af175eec4/volumes" Oct 14 15:50:22 crc kubenswrapper[4945]: I1014 15:50:22.865398 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-qx6hl_d10e063a-f99e-4fc8-b539-224bbf23f466/manager/0.log" Oct 14 15:50:22 crc kubenswrapper[4945]: I1014 15:50:22.923105 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-fl79b_b4e68062-2195-43f9-a6ae-a58d12bd7ce2/kube-rbac-proxy/0.log" Oct 14 15:50:23 crc kubenswrapper[4945]: I1014 15:50:23.006490 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-fl79b_b4e68062-2195-43f9-a6ae-a58d12bd7ce2/manager/0.log" Oct 14 15:50:23 crc kubenswrapper[4945]: I1014 15:50:23.110817 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-vspm4_4af90a70-780d-4d6f-9fc5-6a384002a606/kube-rbac-proxy/0.log" Oct 14 15:50:23 crc kubenswrapper[4945]: I1014 15:50:23.145051 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-vspm4_4af90a70-780d-4d6f-9fc5-6a384002a606/manager/0.log" Oct 14 15:50:23 crc kubenswrapper[4945]: I1014 15:50:23.295382 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w_669db3c1-3b7e-4bbd-980e-93fdea61a1cc/kube-rbac-proxy/0.log" Oct 14 15:50:23 crc kubenswrapper[4945]: I1014 15:50:23.349012 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dflh2w_669db3c1-3b7e-4bbd-980e-93fdea61a1cc/manager/0.log" Oct 14 15:50:23 crc kubenswrapper[4945]: I1014 15:50:23.457445 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-b9d6d9595-9zj7r_26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad/kube-rbac-proxy/0.log" Oct 14 15:50:23 crc kubenswrapper[4945]: I1014 15:50:23.569001 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-674f597f57-cqpn6_795c3628-2ec0-4846-b345-3019636282e9/kube-rbac-proxy/0.log" Oct 14 15:50:23 crc kubenswrapper[4945]: I1014 15:50:23.811951 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-674f597f57-cqpn6_795c3628-2ec0-4846-b345-3019636282e9/operator/0.log" Oct 14 15:50:23 crc kubenswrapper[4945]: I1014 15:50:23.847218 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-f9h5q_4b5016e5-1067-4b63-9db4-f5257354538d/registry-server/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.028647 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-r6tvw_37b893c7-b6a7-4a19-bc3e-b251deacdd27/kube-rbac-proxy/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.096495 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-58vsj_01b34d70-1617-4e4d-81ab-8cada4bb1edf/kube-rbac-proxy/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.109848 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-r6tvw_37b893c7-b6a7-4a19-bc3e-b251deacdd27/manager/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.291012 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-58vsj_01b34d70-1617-4e4d-81ab-8cada4bb1edf/manager/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.320224 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-qcg8q_f23ed77e-a951-4003-bbc2-a2e06431533e/operator/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.345669 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-b9d6d9595-9zj7r_26f6d765-fcf8-4f22-a10b-7e9a06a6c0ad/manager/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: E1014 15:50:24.378090 4945 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a8966c_3784_4d97_9cac_b6a1c4561ad6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a8966c_3784_4d97_9cac_b6a1c4561ad6.slice/crio-0952d34a8b0fe544ed8046ec0396eef638dde6d3d811210c3346511a0db492c4\": RecentStats: unable to find data in memory cache]" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.479232 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-48x4h_e3193e9c-898b-456d-a825-ea8176568803/kube-rbac-proxy/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.500922 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-48x4h_e3193e9c-898b-456d-a825-ea8176568803/manager/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.548059 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-48769_50453e84-c640-4a80-ac7d-0fe2ca24acfd/kube-rbac-proxy/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.635700 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-48769_50453e84-c640-4a80-ac7d-0fe2ca24acfd/manager/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.707288 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-kkqht_2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c/manager/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.711543 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-kkqht_2bd2b6dd-b7a4-4d44-9cb5-5bd89d28918c/kube-rbac-proxy/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.840412 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-sjvtx_2c954bb2-bf12-491f-92e8-637e0eb13ad5/kube-rbac-proxy/0.log" Oct 14 15:50:24 crc kubenswrapper[4945]: I1014 15:50:24.873464 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-sjvtx_2c954bb2-bf12-491f-92e8-637e0eb13ad5/manager/0.log" Oct 14 15:50:30 crc kubenswrapper[4945]: E1014 15:50:30.799067 4945 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/1968cb6a0c21fd896cc734674131229d7ec081c7d952d4c5f6ec5512de2a179d/diff" to get inode usage: stat /var/lib/containers/storage/overlay/1968cb6a0c21fd896cc734674131229d7ec081c7d952d4c5f6ec5512de2a179d/diff: no such file or directory, extraDiskErr: Oct 14 15:50:38 crc kubenswrapper[4945]: I1014 15:50:38.984070 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jphx9_b47bfeba-2bc6-4c04-bd5a-a02b10ef8478/control-plane-machine-set-operator/0.log" Oct 14 15:50:39 crc kubenswrapper[4945]: I1014 15:50:39.134709 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fkbmp_932ab187-513c-4d0f-9c6f-f986ac444f78/kube-rbac-proxy/0.log" Oct 14 15:50:39 crc kubenswrapper[4945]: I1014 15:50:39.142085 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fkbmp_932ab187-513c-4d0f-9c6f-f986ac444f78/machine-api-operator/0.log" Oct 14 15:50:46 crc kubenswrapper[4945]: I1014 15:50:46.795549 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:50:46 crc kubenswrapper[4945]: I1014 15:50:46.796173 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:50:49 crc kubenswrapper[4945]: I1014 15:50:49.391912 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-jmvq2_6a2a312b-2002-4494-b432-efb6060b5a89/cert-manager-controller/0.log" Oct 14 15:50:49 crc kubenswrapper[4945]: I1014 15:50:49.498639 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-kw8kj_2385f0e3-9e8d-4971-95e6-08dbf4af170c/cert-manager-cainjector/0.log" Oct 14 15:50:49 crc kubenswrapper[4945]: I1014 15:50:49.574417 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-j5ssh_a59d5293-e44c-459d-9875-137a289aa9d8/cert-manager-webhook/0.log" Oct 14 15:51:00 crc kubenswrapper[4945]: I1014 15:51:00.084269 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-27p5l_f481cdcc-3f50-4aba-a28a-500e4b19bcff/nmstate-console-plugin/0.log" Oct 14 15:51:00 crc kubenswrapper[4945]: I1014 15:51:00.243453 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-nx8z6_93cf1921-8abd-42e0-8f80-aa5023f63c96/nmstate-handler/0.log" Oct 14 15:51:00 crc kubenswrapper[4945]: I1014 15:51:00.305420 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-dh8w2_97cf21fd-90e2-4591-8563-66afb4145fa4/kube-rbac-proxy/0.log" Oct 14 15:51:00 crc kubenswrapper[4945]: I1014 15:51:00.388841 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-dh8w2_97cf21fd-90e2-4591-8563-66afb4145fa4/nmstate-metrics/0.log" Oct 14 15:51:00 crc kubenswrapper[4945]: I1014 15:51:00.504058 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-pz6t2_27598c0b-8118-406d-b0dd-9e47a23343ad/nmstate-operator/0.log" Oct 14 15:51:00 crc kubenswrapper[4945]: I1014 15:51:00.607722 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-p64sf_76343e33-0b0d-4df5-98cf-4d58247d41f7/nmstate-webhook/0.log" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.514514 4945 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7fft5"] Oct 14 15:51:03 crc kubenswrapper[4945]: E1014 15:51:03.515300 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6049943a-96c2-42d1-9da5-0fdebabafe3a" containerName="extract-content" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.515317 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6049943a-96c2-42d1-9da5-0fdebabafe3a" containerName="extract-content" Oct 14 15:51:03 crc kubenswrapper[4945]: E1014 15:51:03.515337 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6049943a-96c2-42d1-9da5-0fdebabafe3a" containerName="extract-utilities" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.515344 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6049943a-96c2-42d1-9da5-0fdebabafe3a" containerName="extract-utilities" Oct 14 15:51:03 crc kubenswrapper[4945]: E1014 15:51:03.515361 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3811f6a-94ba-4d10-bd4b-873af175eec4" containerName="registry-server" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.515372 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3811f6a-94ba-4d10-bd4b-873af175eec4" containerName="registry-server" Oct 14 15:51:03 crc kubenswrapper[4945]: E1014 15:51:03.515390 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3811f6a-94ba-4d10-bd4b-873af175eec4" containerName="extract-content" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.515397 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3811f6a-94ba-4d10-bd4b-873af175eec4" containerName="extract-content" Oct 14 15:51:03 crc kubenswrapper[4945]: E1014 15:51:03.515421 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6049943a-96c2-42d1-9da5-0fdebabafe3a" containerName="registry-server" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.515428 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="6049943a-96c2-42d1-9da5-0fdebabafe3a" containerName="registry-server" Oct 14 15:51:03 crc kubenswrapper[4945]: E1014 15:51:03.515447 4945 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3811f6a-94ba-4d10-bd4b-873af175eec4" containerName="extract-utilities" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.515455 4945 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3811f6a-94ba-4d10-bd4b-873af175eec4" containerName="extract-utilities" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.515677 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="6049943a-96c2-42d1-9da5-0fdebabafe3a" containerName="registry-server" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.515705 4945 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3811f6a-94ba-4d10-bd4b-873af175eec4" containerName="registry-server" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.517396 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.531864 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7fft5"] Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.611025 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b15f38-b00d-43c1-80b9-b9179f5231f2-catalog-content\") pod \"community-operators-7fft5\" (UID: \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\") " pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.611422 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p99rb\" (UniqueName: \"kubernetes.io/projected/a9b15f38-b00d-43c1-80b9-b9179f5231f2-kube-api-access-p99rb\") pod \"community-operators-7fft5\" (UID: \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\") " pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.611461 4945 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b15f38-b00d-43c1-80b9-b9179f5231f2-utilities\") pod \"community-operators-7fft5\" (UID: \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\") " pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.713028 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p99rb\" (UniqueName: \"kubernetes.io/projected/a9b15f38-b00d-43c1-80b9-b9179f5231f2-kube-api-access-p99rb\") pod \"community-operators-7fft5\" (UID: \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\") " pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.713100 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b15f38-b00d-43c1-80b9-b9179f5231f2-utilities\") pod \"community-operators-7fft5\" (UID: \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\") " pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.713184 4945 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b15f38-b00d-43c1-80b9-b9179f5231f2-catalog-content\") pod \"community-operators-7fft5\" (UID: \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\") " pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.713627 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b15f38-b00d-43c1-80b9-b9179f5231f2-utilities\") pod \"community-operators-7fft5\" (UID: \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\") " pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.713690 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b15f38-b00d-43c1-80b9-b9179f5231f2-catalog-content\") pod \"community-operators-7fft5\" (UID: \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\") " pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.735973 4945 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p99rb\" (UniqueName: \"kubernetes.io/projected/a9b15f38-b00d-43c1-80b9-b9179f5231f2-kube-api-access-p99rb\") pod \"community-operators-7fft5\" (UID: \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\") " pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:03 crc kubenswrapper[4945]: I1014 15:51:03.837008 4945 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:04 crc kubenswrapper[4945]: I1014 15:51:04.422577 4945 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7fft5"] Oct 14 15:51:04 crc kubenswrapper[4945]: I1014 15:51:04.489420 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fft5" event={"ID":"a9b15f38-b00d-43c1-80b9-b9179f5231f2","Type":"ContainerStarted","Data":"24f8b9e07ea539248d3a5a484e09a6737a2570d1fe6188cb774d58b9c7c3368a"} Oct 14 15:51:05 crc kubenswrapper[4945]: I1014 15:51:05.499277 4945 generic.go:334] "Generic (PLEG): container finished" podID="a9b15f38-b00d-43c1-80b9-b9179f5231f2" containerID="c8ae9d43c9e5ad0e6b57e0399c9c1e7f9e76fe63257ce78a9a8ab04b79aca236" exitCode=0 Oct 14 15:51:05 crc kubenswrapper[4945]: I1014 15:51:05.499502 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fft5" event={"ID":"a9b15f38-b00d-43c1-80b9-b9179f5231f2","Type":"ContainerDied","Data":"c8ae9d43c9e5ad0e6b57e0399c9c1e7f9e76fe63257ce78a9a8ab04b79aca236"} Oct 14 15:51:07 crc kubenswrapper[4945]: I1014 15:51:07.521360 4945 generic.go:334] "Generic (PLEG): container finished" podID="a9b15f38-b00d-43c1-80b9-b9179f5231f2" containerID="702f2e301bbae9f4a2e44171ebffc6534d92a0c1c63661b6cfcff379f07e7cc3" exitCode=0 Oct 14 15:51:07 crc kubenswrapper[4945]: I1014 15:51:07.521827 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fft5" event={"ID":"a9b15f38-b00d-43c1-80b9-b9179f5231f2","Type":"ContainerDied","Data":"702f2e301bbae9f4a2e44171ebffc6534d92a0c1c63661b6cfcff379f07e7cc3"} Oct 14 15:51:08 crc kubenswrapper[4945]: I1014 15:51:08.533405 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fft5" event={"ID":"a9b15f38-b00d-43c1-80b9-b9179f5231f2","Type":"ContainerStarted","Data":"542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1"} Oct 14 15:51:08 crc kubenswrapper[4945]: I1014 15:51:08.555137 4945 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7fft5" podStartSLOduration=2.9152030399999997 podStartE2EDuration="5.555114554s" podCreationTimestamp="2025-10-14 15:51:03 +0000 UTC" firstStartedPulling="2025-10-14 15:51:05.502283908 +0000 UTC m=+2195.486332276" lastFinishedPulling="2025-10-14 15:51:08.142195422 +0000 UTC m=+2198.126243790" observedRunningTime="2025-10-14 15:51:08.551381847 +0000 UTC m=+2198.535430225" watchObservedRunningTime="2025-10-14 15:51:08.555114554 +0000 UTC m=+2198.539162922" Oct 14 15:51:13 crc kubenswrapper[4945]: I1014 15:51:13.838057 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:13 crc kubenswrapper[4945]: I1014 15:51:13.838567 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:13 crc kubenswrapper[4945]: I1014 15:51:13.888016 4945 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:13 crc kubenswrapper[4945]: I1014 15:51:13.998067 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-tvn99_38c1e540-b577-472f-82e3-5079b3823d87/kube-rbac-proxy/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.116096 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-tvn99_38c1e540-b577-472f-82e3-5079b3823d87/controller/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.178988 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-frr-files/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.341838 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-frr-files/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.346155 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-metrics/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.346494 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-reloader/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.364677 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-reloader/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.539534 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-frr-files/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.555444 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-reloader/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.606787 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-metrics/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.607491 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-metrics/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.646620 4945 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.699709 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7fft5"] Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.803695 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-frr-files/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.856890 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-reloader/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.887126 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/cp-metrics/0.log" Oct 14 15:51:14 crc kubenswrapper[4945]: I1014 15:51:14.890291 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/controller/0.log" Oct 14 15:51:15 crc kubenswrapper[4945]: I1014 15:51:15.070462 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/frr-metrics/0.log" Oct 14 15:51:15 crc kubenswrapper[4945]: I1014 15:51:15.071738 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/kube-rbac-proxy/0.log" Oct 14 15:51:15 crc kubenswrapper[4945]: I1014 15:51:15.111868 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/kube-rbac-proxy-frr/0.log" Oct 14 15:51:15 crc kubenswrapper[4945]: I1014 15:51:15.262563 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/reloader/0.log" Oct 14 15:51:15 crc kubenswrapper[4945]: I1014 15:51:15.353816 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-pfbt9_79b86a9d-a49c-428a-b3f9-3e2625ad1abc/frr-k8s-webhook-server/0.log" Oct 14 15:51:15 crc kubenswrapper[4945]: I1014 15:51:15.593074 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9f47f9675-47zvd_740d5a8e-2e4a-4fd1-8412-d5b989734f2f/manager/0.log" Oct 14 15:51:15 crc kubenswrapper[4945]: I1014 15:51:15.753835 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6545bcf4ff-f248w_59c45798-dcb8-4fae-ba47-82a861d97023/webhook-server/0.log" Oct 14 15:51:15 crc kubenswrapper[4945]: I1014 15:51:15.834512 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-lgvg5_9a07810a-86fd-4dd6-a704-037bea9afcab/kube-rbac-proxy/0.log" Oct 14 15:51:16 crc kubenswrapper[4945]: I1014 15:51:16.049367 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h88zg_2b3404d8-44ad-4e9c-9c35-1d65d401b9cb/frr/0.log" Oct 14 15:51:16 crc kubenswrapper[4945]: I1014 15:51:16.281325 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-lgvg5_9a07810a-86fd-4dd6-a704-037bea9afcab/speaker/0.log" Oct 14 15:51:16 crc kubenswrapper[4945]: I1014 15:51:16.603305 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7fft5" podUID="a9b15f38-b00d-43c1-80b9-b9179f5231f2" containerName="registry-server" containerID="cri-o://542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1" gracePeriod=2 Oct 14 15:51:16 crc kubenswrapper[4945]: I1014 15:51:16.795454 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:51:16 crc kubenswrapper[4945]: I1014 15:51:16.795773 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:51:16 crc kubenswrapper[4945]: I1014 15:51:16.795912 4945 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" Oct 14 15:51:16 crc kubenswrapper[4945]: I1014 15:51:16.796657 4945 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0a877f7512be56a61764de5bfd6cb2d36289868e677eb3453925231a2256b2a9"} pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 15:51:16 crc kubenswrapper[4945]: I1014 15:51:16.796728 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" containerID="cri-o://0a877f7512be56a61764de5bfd6cb2d36289868e677eb3453925231a2256b2a9" gracePeriod=600 Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.132569 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.271680 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b15f38-b00d-43c1-80b9-b9179f5231f2-utilities\") pod \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\" (UID: \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\") " Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.271768 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b15f38-b00d-43c1-80b9-b9179f5231f2-catalog-content\") pod \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\" (UID: \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\") " Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.271961 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p99rb\" (UniqueName: \"kubernetes.io/projected/a9b15f38-b00d-43c1-80b9-b9179f5231f2-kube-api-access-p99rb\") pod \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\" (UID: \"a9b15f38-b00d-43c1-80b9-b9179f5231f2\") " Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.272946 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9b15f38-b00d-43c1-80b9-b9179f5231f2-utilities" (OuterVolumeSpecName: "utilities") pod "a9b15f38-b00d-43c1-80b9-b9179f5231f2" (UID: "a9b15f38-b00d-43c1-80b9-b9179f5231f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.288909 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9b15f38-b00d-43c1-80b9-b9179f5231f2-kube-api-access-p99rb" (OuterVolumeSpecName: "kube-api-access-p99rb") pod "a9b15f38-b00d-43c1-80b9-b9179f5231f2" (UID: "a9b15f38-b00d-43c1-80b9-b9179f5231f2"). InnerVolumeSpecName "kube-api-access-p99rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.326687 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9b15f38-b00d-43c1-80b9-b9179f5231f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9b15f38-b00d-43c1-80b9-b9179f5231f2" (UID: "a9b15f38-b00d-43c1-80b9-b9179f5231f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.373793 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p99rb\" (UniqueName: \"kubernetes.io/projected/a9b15f38-b00d-43c1-80b9-b9179f5231f2-kube-api-access-p99rb\") on node \"crc\" DevicePath \"\"" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.373825 4945 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b15f38-b00d-43c1-80b9-b9179f5231f2-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.373839 4945 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b15f38-b00d-43c1-80b9-b9179f5231f2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.613511 4945 generic.go:334] "Generic (PLEG): container finished" podID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerID="0a877f7512be56a61764de5bfd6cb2d36289868e677eb3453925231a2256b2a9" exitCode=0 Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.613570 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerDied","Data":"0a877f7512be56a61764de5bfd6cb2d36289868e677eb3453925231a2256b2a9"} Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.613602 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" event={"ID":"08049862-1ea7-4efc-b9ed-df61b47d430b","Type":"ContainerStarted","Data":"4aa2f710215a4310bfe9434183f32a8bd59b1e9c3fb975022e2a3efcb0f44762"} Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.613619 4945 scope.go:117] "RemoveContainer" containerID="bf5ec5e4fd47e09e38a95edf07e997a98cb69d4581073fc0c86609c6cba807a3" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.617523 4945 generic.go:334] "Generic (PLEG): container finished" podID="a9b15f38-b00d-43c1-80b9-b9179f5231f2" containerID="542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1" exitCode=0 Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.617564 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fft5" event={"ID":"a9b15f38-b00d-43c1-80b9-b9179f5231f2","Type":"ContainerDied","Data":"542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1"} Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.617595 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fft5" event={"ID":"a9b15f38-b00d-43c1-80b9-b9179f5231f2","Type":"ContainerDied","Data":"24f8b9e07ea539248d3a5a484e09a6737a2570d1fe6188cb774d58b9c7c3368a"} Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.617598 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fft5" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.666187 4945 scope.go:117] "RemoveContainer" containerID="542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.672001 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7fft5"] Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.679649 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7fft5"] Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.692438 4945 scope.go:117] "RemoveContainer" containerID="702f2e301bbae9f4a2e44171ebffc6534d92a0c1c63661b6cfcff379f07e7cc3" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.716861 4945 scope.go:117] "RemoveContainer" containerID="c8ae9d43c9e5ad0e6b57e0399c9c1e7f9e76fe63257ce78a9a8ab04b79aca236" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.774574 4945 scope.go:117] "RemoveContainer" containerID="542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1" Oct 14 15:51:17 crc kubenswrapper[4945]: E1014 15:51:17.775317 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1\": container with ID starting with 542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1 not found: ID does not exist" containerID="542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.775439 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1"} err="failed to get container status \"542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1\": rpc error: code = NotFound desc = could not find container \"542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1\": container with ID starting with 542186240f0198855f3861aacbfb3a218a3d476d5f3946f8c18e513515d886c1 not found: ID does not exist" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.775527 4945 scope.go:117] "RemoveContainer" containerID="702f2e301bbae9f4a2e44171ebffc6534d92a0c1c63661b6cfcff379f07e7cc3" Oct 14 15:51:17 crc kubenswrapper[4945]: E1014 15:51:17.778458 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"702f2e301bbae9f4a2e44171ebffc6534d92a0c1c63661b6cfcff379f07e7cc3\": container with ID starting with 702f2e301bbae9f4a2e44171ebffc6534d92a0c1c63661b6cfcff379f07e7cc3 not found: ID does not exist" containerID="702f2e301bbae9f4a2e44171ebffc6534d92a0c1c63661b6cfcff379f07e7cc3" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.778591 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"702f2e301bbae9f4a2e44171ebffc6534d92a0c1c63661b6cfcff379f07e7cc3"} err="failed to get container status \"702f2e301bbae9f4a2e44171ebffc6534d92a0c1c63661b6cfcff379f07e7cc3\": rpc error: code = NotFound desc = could not find container \"702f2e301bbae9f4a2e44171ebffc6534d92a0c1c63661b6cfcff379f07e7cc3\": container with ID starting with 702f2e301bbae9f4a2e44171ebffc6534d92a0c1c63661b6cfcff379f07e7cc3 not found: ID does not exist" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.778684 4945 scope.go:117] "RemoveContainer" containerID="c8ae9d43c9e5ad0e6b57e0399c9c1e7f9e76fe63257ce78a9a8ab04b79aca236" Oct 14 15:51:17 crc kubenswrapper[4945]: E1014 15:51:17.779263 4945 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8ae9d43c9e5ad0e6b57e0399c9c1e7f9e76fe63257ce78a9a8ab04b79aca236\": container with ID starting with c8ae9d43c9e5ad0e6b57e0399c9c1e7f9e76fe63257ce78a9a8ab04b79aca236 not found: ID does not exist" containerID="c8ae9d43c9e5ad0e6b57e0399c9c1e7f9e76fe63257ce78a9a8ab04b79aca236" Oct 14 15:51:17 crc kubenswrapper[4945]: I1014 15:51:17.779377 4945 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8ae9d43c9e5ad0e6b57e0399c9c1e7f9e76fe63257ce78a9a8ab04b79aca236"} err="failed to get container status \"c8ae9d43c9e5ad0e6b57e0399c9c1e7f9e76fe63257ce78a9a8ab04b79aca236\": rpc error: code = NotFound desc = could not find container \"c8ae9d43c9e5ad0e6b57e0399c9c1e7f9e76fe63257ce78a9a8ab04b79aca236\": container with ID starting with c8ae9d43c9e5ad0e6b57e0399c9c1e7f9e76fe63257ce78a9a8ab04b79aca236 not found: ID does not exist" Oct 14 15:51:18 crc kubenswrapper[4945]: I1014 15:51:18.774682 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9b15f38-b00d-43c1-80b9-b9179f5231f2" path="/var/lib/kubelet/pods/a9b15f38-b00d-43c1-80b9-b9179f5231f2/volumes" Oct 14 15:51:27 crc kubenswrapper[4945]: I1014 15:51:27.314832 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/util/0.log" Oct 14 15:51:27 crc kubenswrapper[4945]: I1014 15:51:27.491349 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/pull/0.log" Oct 14 15:51:27 crc kubenswrapper[4945]: I1014 15:51:27.502977 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/pull/0.log" Oct 14 15:51:27 crc kubenswrapper[4945]: I1014 15:51:27.523314 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/util/0.log" Oct 14 15:51:27 crc kubenswrapper[4945]: I1014 15:51:27.676447 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/util/0.log" Oct 14 15:51:27 crc kubenswrapper[4945]: I1014 15:51:27.697629 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/pull/0.log" Oct 14 15:51:27 crc kubenswrapper[4945]: I1014 15:51:27.731729 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2x5xwp_c73879b2-ec9f-4bb0-8b7c-6ec24f36be5d/extract/0.log" Oct 14 15:51:27 crc kubenswrapper[4945]: I1014 15:51:27.856094 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8kt6g_9de0e862-25dd-46e1-9173-2419826478a8/extract-utilities/0.log" Oct 14 15:51:28 crc kubenswrapper[4945]: I1014 15:51:28.024865 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8kt6g_9de0e862-25dd-46e1-9173-2419826478a8/extract-utilities/0.log" Oct 14 15:51:28 crc kubenswrapper[4945]: I1014 15:51:28.026246 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8kt6g_9de0e862-25dd-46e1-9173-2419826478a8/extract-content/0.log" Oct 14 15:51:28 crc kubenswrapper[4945]: I1014 15:51:28.056843 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8kt6g_9de0e862-25dd-46e1-9173-2419826478a8/extract-content/0.log" Oct 14 15:51:28 crc kubenswrapper[4945]: I1014 15:51:28.198749 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8kt6g_9de0e862-25dd-46e1-9173-2419826478a8/extract-utilities/0.log" Oct 14 15:51:28 crc kubenswrapper[4945]: I1014 15:51:28.248523 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8kt6g_9de0e862-25dd-46e1-9173-2419826478a8/extract-content/0.log" Oct 14 15:51:28 crc kubenswrapper[4945]: I1014 15:51:28.352293 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8kt6g_9de0e862-25dd-46e1-9173-2419826478a8/registry-server/0.log" Oct 14 15:51:28 crc kubenswrapper[4945]: I1014 15:51:28.406855 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/extract-utilities/0.log" Oct 14 15:51:28 crc kubenswrapper[4945]: I1014 15:51:28.644372 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/extract-content/0.log" Oct 14 15:51:28 crc kubenswrapper[4945]: I1014 15:51:28.655070 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/extract-utilities/0.log" Oct 14 15:51:28 crc kubenswrapper[4945]: I1014 15:51:28.676713 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/extract-content/0.log" Oct 14 15:51:28 crc kubenswrapper[4945]: I1014 15:51:28.802936 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/extract-content/0.log" Oct 14 15:51:28 crc kubenswrapper[4945]: I1014 15:51:28.817005 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/extract-utilities/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.013300 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/util/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.204083 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/util/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.227887 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/pull/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.277367 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/pull/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.306674 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4cn2_63cb59cc-d19f-41e4-a839-1b67f86a5ffe/registry-server/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.394297 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/util/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.411054 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/pull/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.431840 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4kd8c_a90d4b1f-7443-474c-b4b7-7cb7ac9ef398/extract/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.633483 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hdrgb_1139d59d-b224-467d-849e-af47cba636e6/marketplace-operator/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.649323 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/extract-utilities/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.781939 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/extract-utilities/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.814440 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/extract-content/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.831275 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/extract-content/0.log" Oct 14 15:51:29 crc kubenswrapper[4945]: I1014 15:51:29.991066 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/extract-utilities/0.log" Oct 14 15:51:30 crc kubenswrapper[4945]: I1014 15:51:30.015544 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/extract-content/0.log" Oct 14 15:51:30 crc kubenswrapper[4945]: I1014 15:51:30.100985 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9lt9_0f2803bc-a79f-4af4-99c7-f372b3b07bc8/registry-server/0.log" Oct 14 15:51:30 crc kubenswrapper[4945]: I1014 15:51:30.176012 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/extract-utilities/0.log" Oct 14 15:51:30 crc kubenswrapper[4945]: I1014 15:51:30.332676 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/extract-content/0.log" Oct 14 15:51:30 crc kubenswrapper[4945]: I1014 15:51:30.341096 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/extract-utilities/0.log" Oct 14 15:51:30 crc kubenswrapper[4945]: I1014 15:51:30.361860 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/extract-content/0.log" Oct 14 15:51:30 crc kubenswrapper[4945]: I1014 15:51:30.532676 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/extract-utilities/0.log" Oct 14 15:51:30 crc kubenswrapper[4945]: I1014 15:51:30.644392 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/extract-content/0.log" Oct 14 15:51:31 crc kubenswrapper[4945]: I1014 15:51:31.047252 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-s9z56_33acd0ef-71bb-467c-be47-0c93a694c8f2/registry-server/0.log" Oct 14 15:52:59 crc kubenswrapper[4945]: I1014 15:52:59.508590 4945 generic.go:334] "Generic (PLEG): container finished" podID="6eaf1e4a-64f3-4cba-a83d-900ba092249c" containerID="6e65a24520ae1bdd6636af8e8c55bc0923d07f1b61baeb2586d875fde80306a1" exitCode=0 Oct 14 15:52:59 crc kubenswrapper[4945]: I1014 15:52:59.508699 4945 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6v2cl/must-gather-tw6jh" event={"ID":"6eaf1e4a-64f3-4cba-a83d-900ba092249c","Type":"ContainerDied","Data":"6e65a24520ae1bdd6636af8e8c55bc0923d07f1b61baeb2586d875fde80306a1"} Oct 14 15:52:59 crc kubenswrapper[4945]: I1014 15:52:59.509940 4945 scope.go:117] "RemoveContainer" containerID="6e65a24520ae1bdd6636af8e8c55bc0923d07f1b61baeb2586d875fde80306a1" Oct 14 15:53:00 crc kubenswrapper[4945]: I1014 15:53:00.223081 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6v2cl_must-gather-tw6jh_6eaf1e4a-64f3-4cba-a83d-900ba092249c/gather/0.log" Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.187247 4945 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6v2cl/must-gather-tw6jh"] Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.187997 4945 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-6v2cl/must-gather-tw6jh" podUID="6eaf1e4a-64f3-4cba-a83d-900ba092249c" containerName="copy" containerID="cri-o://5d62a6359df9872e4c89fbdca62fe2f9cf9a771a978df806b9e03c91ea3dc72e" gracePeriod=2 Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.195912 4945 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6v2cl/must-gather-tw6jh"] Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.604495 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6v2cl_must-gather-tw6jh_6eaf1e4a-64f3-4cba-a83d-900ba092249c/copy/0.log" Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.605260 4945 generic.go:334] "Generic (PLEG): container finished" podID="6eaf1e4a-64f3-4cba-a83d-900ba092249c" containerID="5d62a6359df9872e4c89fbdca62fe2f9cf9a771a978df806b9e03c91ea3dc72e" exitCode=143 Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.605319 4945 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02be56e4602d9fe551a84bb38428d65b85e7a50857f220ce15372569649aed51" Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.626965 4945 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6v2cl_must-gather-tw6jh_6eaf1e4a-64f3-4cba-a83d-900ba092249c/copy/0.log" Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.627761 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/must-gather-tw6jh" Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.722021 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6eaf1e4a-64f3-4cba-a83d-900ba092249c-must-gather-output\") pod \"6eaf1e4a-64f3-4cba-a83d-900ba092249c\" (UID: \"6eaf1e4a-64f3-4cba-a83d-900ba092249c\") " Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.722142 4945 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqdpj\" (UniqueName: \"kubernetes.io/projected/6eaf1e4a-64f3-4cba-a83d-900ba092249c-kube-api-access-kqdpj\") pod \"6eaf1e4a-64f3-4cba-a83d-900ba092249c\" (UID: \"6eaf1e4a-64f3-4cba-a83d-900ba092249c\") " Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.742443 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eaf1e4a-64f3-4cba-a83d-900ba092249c-kube-api-access-kqdpj" (OuterVolumeSpecName: "kube-api-access-kqdpj") pod "6eaf1e4a-64f3-4cba-a83d-900ba092249c" (UID: "6eaf1e4a-64f3-4cba-a83d-900ba092249c"). InnerVolumeSpecName "kube-api-access-kqdpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.825049 4945 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqdpj\" (UniqueName: \"kubernetes.io/projected/6eaf1e4a-64f3-4cba-a83d-900ba092249c-kube-api-access-kqdpj\") on node \"crc\" DevicePath \"\"" Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.872133 4945 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eaf1e4a-64f3-4cba-a83d-900ba092249c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6eaf1e4a-64f3-4cba-a83d-900ba092249c" (UID: "6eaf1e4a-64f3-4cba-a83d-900ba092249c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 15:53:10 crc kubenswrapper[4945]: I1014 15:53:10.926575 4945 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6eaf1e4a-64f3-4cba-a83d-900ba092249c-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 14 15:53:11 crc kubenswrapper[4945]: I1014 15:53:11.623538 4945 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6v2cl/must-gather-tw6jh" Oct 14 15:53:12 crc kubenswrapper[4945]: I1014 15:53:12.775148 4945 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eaf1e4a-64f3-4cba-a83d-900ba092249c" path="/var/lib/kubelet/pods/6eaf1e4a-64f3-4cba-a83d-900ba092249c/volumes" Oct 14 15:53:46 crc kubenswrapper[4945]: I1014 15:53:46.795469 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:53:46 crc kubenswrapper[4945]: I1014 15:53:46.796050 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 15:54:16 crc kubenswrapper[4945]: I1014 15:54:16.795697 4945 patch_prober.go:28] interesting pod/machine-config-daemon-tx7tr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 15:54:16 crc kubenswrapper[4945]: I1014 15:54:16.796334 4945 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tx7tr" podUID="08049862-1ea7-4efc-b9ed-df61b47d430b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515073471104024447 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015073471105017365 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015073464033016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015073464033015462 5ustar corecore